00:00:00.000 Started by upstream project "autotest-per-patch" build number 126212 00:00:00.000 originally caused by: 00:00:00.000 Started by user sys_sgci 00:00:00.095 Checking out git https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool into /var/jenkins_home/workspace/nvmf-tcp-phy-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4 to read jbp/jenkins/jjb-config/jobs/autotest-downstream/autotest-phy.groovy 00:00:00.096 The recommended git tool is: git 00:00:00.096 using credential 00000000-0000-0000-0000-000000000002 00:00:00.098 > git rev-parse --resolve-git-dir /var/jenkins_home/workspace/nvmf-tcp-phy-autotest_script/33b20b30f0a51e6b52980845e0f6aa336787973ad45e341fbbf98d1b65b265d4/jbp/.git # timeout=10 00:00:00.134 Fetching changes from the remote Git repository 00:00:00.137 > git config remote.origin.url https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool # timeout=10 00:00:00.170 Using shallow fetch with depth 1 00:00:00.170 Fetching upstream changes from https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool 00:00:00.170 > git --version # timeout=10 00:00:00.200 > git --version # 'git version 2.39.2' 00:00:00.200 using GIT_ASKPASS to set credentials SPDKCI HTTPS Credentials 00:00:00.224 Setting http proxy: proxy-dmz.intel.com:911 00:00:00.224 > git fetch --tags --force --progress --depth=1 -- https://review.spdk.io/gerrit/a/build_pool/jenkins_build_pool refs/heads/master # timeout=5 00:00:05.333 > git rev-parse origin/FETCH_HEAD^{commit} # timeout=10 00:00:05.346 > git rev-parse FETCH_HEAD^{commit} # timeout=10 00:00:05.359 Checking out Revision 7caca6989ac753a10259529aadac5754060382af (FETCH_HEAD) 00:00:05.359 > git config core.sparsecheckout # timeout=10 00:00:05.371 > git read-tree -mu HEAD # timeout=10 00:00:05.390 > git checkout -f 7caca6989ac753a10259529aadac5754060382af # timeout=5 00:00:05.410 Commit message: "jenkins/jjb-config: Purge centos leftovers" 00:00:05.410 > git rev-list --no-walk 7caca6989ac753a10259529aadac5754060382af # timeout=10 00:00:05.495 [Pipeline] Start of Pipeline 00:00:05.512 [Pipeline] library 00:00:05.515 Loading library shm_lib@master 00:00:05.515 Library shm_lib@master is cached. Copying from home. 00:00:05.534 [Pipeline] node 00:00:05.544 Running on WFP8 in /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:00:05.545 [Pipeline] { 00:00:05.556 [Pipeline] catchError 00:00:05.557 [Pipeline] { 00:00:05.572 [Pipeline] wrap 00:00:05.582 [Pipeline] { 00:00:05.589 [Pipeline] stage 00:00:05.590 [Pipeline] { (Prologue) 00:00:05.734 [Pipeline] sh 00:00:06.019 + logger -p user.info -t JENKINS-CI 00:00:06.038 [Pipeline] echo 00:00:06.039 Node: WFP8 00:00:06.046 [Pipeline] sh 00:00:06.347 [Pipeline] setCustomBuildProperty 00:00:06.357 [Pipeline] echo 00:00:06.358 Cleanup processes 00:00:06.362 [Pipeline] sh 00:00:06.642 + sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:06.642 297639 sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:06.659 [Pipeline] sh 00:00:06.946 ++ sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:06.946 ++ grep -v 'sudo pgrep' 00:00:06.946 ++ awk '{print $1}' 00:00:06.946 + sudo kill -9 00:00:06.946 + true 00:00:06.963 [Pipeline] cleanWs 00:00:06.973 [WS-CLEANUP] Deleting project workspace... 00:00:06.973 [WS-CLEANUP] Deferred wipeout is used... 00:00:06.980 [WS-CLEANUP] done 00:00:06.985 [Pipeline] setCustomBuildProperty 00:00:06.999 [Pipeline] sh 00:00:07.290 + sudo git config --global --replace-all safe.directory '*' 00:00:07.424 [Pipeline] httpRequest 00:00:07.453 [Pipeline] echo 00:00:07.455 Sorcerer 10.211.164.101 is alive 00:00:07.461 [Pipeline] httpRequest 00:00:07.465 HttpMethod: GET 00:00:07.465 URL: http://10.211.164.101/packages/jbp_7caca6989ac753a10259529aadac5754060382af.tar.gz 00:00:07.466 Sending request to url: http://10.211.164.101/packages/jbp_7caca6989ac753a10259529aadac5754060382af.tar.gz 00:00:07.485 Response Code: HTTP/1.1 200 OK 00:00:07.485 Success: Status code 200 is in the accepted range: 200,404 00:00:07.486 Saving response body to /var/jenkins/workspace/nvmf-tcp-phy-autotest/jbp_7caca6989ac753a10259529aadac5754060382af.tar.gz 00:00:12.011 [Pipeline] sh 00:00:12.293 + tar --no-same-owner -xf jbp_7caca6989ac753a10259529aadac5754060382af.tar.gz 00:00:12.309 [Pipeline] httpRequest 00:00:12.327 [Pipeline] echo 00:00:12.329 Sorcerer 10.211.164.101 is alive 00:00:12.336 [Pipeline] httpRequest 00:00:12.340 HttpMethod: GET 00:00:12.340 URL: http://10.211.164.101/packages/spdk_2f3522da79be0b4b631aa0ad68765970a588003a.tar.gz 00:00:12.340 Sending request to url: http://10.211.164.101/packages/spdk_2f3522da79be0b4b631aa0ad68765970a588003a.tar.gz 00:00:12.360 Response Code: HTTP/1.1 200 OK 00:00:12.360 Success: Status code 200 is in the accepted range: 200,404 00:00:12.361 Saving response body to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk_2f3522da79be0b4b631aa0ad68765970a588003a.tar.gz 00:00:49.543 [Pipeline] sh 00:00:49.831 + tar --no-same-owner -xf spdk_2f3522da79be0b4b631aa0ad68765970a588003a.tar.gz 00:00:52.380 [Pipeline] sh 00:00:52.663 + git -C spdk log --oneline -n5 00:00:52.664 2f3522da7 nvmf: move register nvmf_poll_group_poll interrupt to nvmf 00:00:52.664 ef59a6f4b nvmf/tcp: replace pending_buf_queue with nvmf_tcp_request_get_buffers 00:00:52.664 a26f69189 nvmf: enable iobuf based queuing for nvmf requests 00:00:52.664 24034319f nvmf/tcp: use sock group polling for the listening sockets 00:00:52.664 245333351 nvmf/tcp: add transport field to the spdk_nvmf_tcp_port struct 00:00:52.676 [Pipeline] } 00:00:52.694 [Pipeline] // stage 00:00:52.705 [Pipeline] stage 00:00:52.708 [Pipeline] { (Prepare) 00:00:52.728 [Pipeline] writeFile 00:00:52.746 [Pipeline] sh 00:00:53.031 + logger -p user.info -t JENKINS-CI 00:00:53.044 [Pipeline] sh 00:00:53.328 + logger -p user.info -t JENKINS-CI 00:00:53.377 [Pipeline] sh 00:00:53.660 + cat autorun-spdk.conf 00:00:53.661 SPDK_RUN_FUNCTIONAL_TEST=1 00:00:53.661 SPDK_TEST_NVMF=1 00:00:53.661 SPDK_TEST_NVME_CLI=1 00:00:53.661 SPDK_TEST_NVMF_TRANSPORT=tcp 00:00:53.661 SPDK_TEST_NVMF_NICS=e810 00:00:53.661 SPDK_TEST_VFIOUSER=1 00:00:53.661 SPDK_RUN_UBSAN=1 00:00:53.661 NET_TYPE=phy 00:00:53.668 RUN_NIGHTLY=0 00:00:53.674 [Pipeline] readFile 00:00:53.702 [Pipeline] withEnv 00:00:53.705 [Pipeline] { 00:00:53.719 [Pipeline] sh 00:00:54.005 + set -ex 00:00:54.005 + [[ -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf ]] 00:00:54.005 + source /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf 00:00:54.005 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:00:54.005 ++ SPDK_TEST_NVMF=1 00:00:54.005 ++ SPDK_TEST_NVME_CLI=1 00:00:54.005 ++ SPDK_TEST_NVMF_TRANSPORT=tcp 00:00:54.005 ++ SPDK_TEST_NVMF_NICS=e810 00:00:54.005 ++ SPDK_TEST_VFIOUSER=1 00:00:54.005 ++ SPDK_RUN_UBSAN=1 00:00:54.005 ++ NET_TYPE=phy 00:00:54.005 ++ RUN_NIGHTLY=0 00:00:54.005 + case $SPDK_TEST_NVMF_NICS in 00:00:54.005 + DRIVERS=ice 00:00:54.005 + [[ tcp == \r\d\m\a ]] 00:00:54.005 + [[ -n ice ]] 00:00:54.005 + sudo rmmod mlx4_ib mlx5_ib irdma i40iw iw_cxgb4 00:00:54.005 rmmod: ERROR: Module mlx4_ib is not currently loaded 00:00:54.005 rmmod: ERROR: Module mlx5_ib is not currently loaded 00:00:54.005 rmmod: ERROR: Module irdma is not currently loaded 00:00:54.005 rmmod: ERROR: Module i40iw is not currently loaded 00:00:54.005 rmmod: ERROR: Module iw_cxgb4 is not currently loaded 00:00:54.005 + true 00:00:54.005 + for D in $DRIVERS 00:00:54.005 + sudo modprobe ice 00:00:54.005 + exit 0 00:00:54.014 [Pipeline] } 00:00:54.032 [Pipeline] // withEnv 00:00:54.038 [Pipeline] } 00:00:54.053 [Pipeline] // stage 00:00:54.063 [Pipeline] catchError 00:00:54.065 [Pipeline] { 00:00:54.078 [Pipeline] timeout 00:00:54.078 Timeout set to expire in 50 min 00:00:54.081 [Pipeline] { 00:00:54.093 [Pipeline] stage 00:00:54.094 [Pipeline] { (Tests) 00:00:54.107 [Pipeline] sh 00:00:54.386 + jbp/jenkins/jjb-config/jobs/scripts/autoruner.sh /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:00:54.386 ++ readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:00:54.386 + DIR_ROOT=/var/jenkins/workspace/nvmf-tcp-phy-autotest 00:00:54.386 + [[ -n /var/jenkins/workspace/nvmf-tcp-phy-autotest ]] 00:00:54.386 + DIR_SPDK=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:54.386 + DIR_OUTPUT=/var/jenkins/workspace/nvmf-tcp-phy-autotest/output 00:00:54.386 + [[ -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk ]] 00:00:54.386 + [[ ! -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/output ]] 00:00:54.386 + mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/output 00:00:54.386 + [[ -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/output ]] 00:00:54.386 + [[ nvmf-tcp-phy-autotest == pkgdep-* ]] 00:00:54.386 + cd /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:00:54.386 + source /etc/os-release 00:00:54.386 ++ NAME='Fedora Linux' 00:00:54.386 ++ VERSION='38 (Cloud Edition)' 00:00:54.386 ++ ID=fedora 00:00:54.386 ++ VERSION_ID=38 00:00:54.386 ++ VERSION_CODENAME= 00:00:54.386 ++ PLATFORM_ID=platform:f38 00:00:54.386 ++ PRETTY_NAME='Fedora Linux 38 (Cloud Edition)' 00:00:54.386 ++ ANSI_COLOR='0;38;2;60;110;180' 00:00:54.386 ++ LOGO=fedora-logo-icon 00:00:54.386 ++ CPE_NAME=cpe:/o:fedoraproject:fedora:38 00:00:54.386 ++ HOME_URL=https://fedoraproject.org/ 00:00:54.386 ++ DOCUMENTATION_URL=https://docs.fedoraproject.org/en-US/fedora/f38/system-administrators-guide/ 00:00:54.386 ++ SUPPORT_URL=https://ask.fedoraproject.org/ 00:00:54.386 ++ BUG_REPORT_URL=https://bugzilla.redhat.com/ 00:00:54.386 ++ REDHAT_BUGZILLA_PRODUCT=Fedora 00:00:54.386 ++ REDHAT_BUGZILLA_PRODUCT_VERSION=38 00:00:54.386 ++ REDHAT_SUPPORT_PRODUCT=Fedora 00:00:54.386 ++ REDHAT_SUPPORT_PRODUCT_VERSION=38 00:00:54.386 ++ SUPPORT_END=2024-05-14 00:00:54.386 ++ VARIANT='Cloud Edition' 00:00:54.386 ++ VARIANT_ID=cloud 00:00:54.386 + uname -a 00:00:54.386 Linux spdk-wfp-08 6.7.0-68.fc38.x86_64 #1 SMP PREEMPT_DYNAMIC Mon Jan 15 00:59:40 UTC 2024 x86_64 GNU/Linux 00:00:54.386 + sudo /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh status 00:00:56.293 Hugepages 00:00:56.293 node hugesize free / total 00:00:56.293 node0 1048576kB 0 / 0 00:00:56.293 node0 2048kB 0 / 0 00:00:56.293 node1 1048576kB 0 / 0 00:00:56.293 node1 2048kB 0 / 0 00:00:56.293 00:00:56.293 Type BDF Vendor Device NUMA Driver Device Block devices 00:00:56.293 I/OAT 0000:00:04.0 8086 2021 0 ioatdma - - 00:00:56.293 I/OAT 0000:00:04.1 8086 2021 0 ioatdma - - 00:00:56.293 I/OAT 0000:00:04.2 8086 2021 0 ioatdma - - 00:00:56.293 I/OAT 0000:00:04.3 8086 2021 0 ioatdma - - 00:00:56.293 I/OAT 0000:00:04.4 8086 2021 0 ioatdma - - 00:00:56.293 I/OAT 0000:00:04.5 8086 2021 0 ioatdma - - 00:00:56.293 I/OAT 0000:00:04.6 8086 2021 0 ioatdma - - 00:00:56.293 I/OAT 0000:00:04.7 8086 2021 0 ioatdma - - 00:00:56.293 NVMe 0000:5e:00.0 8086 0a54 0 nvme nvme0 nvme0n1 00:00:56.293 I/OAT 0000:80:04.0 8086 2021 1 ioatdma - - 00:00:56.293 I/OAT 0000:80:04.1 8086 2021 1 ioatdma - - 00:00:56.293 I/OAT 0000:80:04.2 8086 2021 1 ioatdma - - 00:00:56.553 I/OAT 0000:80:04.3 8086 2021 1 ioatdma - - 00:00:56.553 I/OAT 0000:80:04.4 8086 2021 1 ioatdma - - 00:00:56.553 I/OAT 0000:80:04.5 8086 2021 1 ioatdma - - 00:00:56.553 I/OAT 0000:80:04.6 8086 2021 1 ioatdma - - 00:00:56.553 I/OAT 0000:80:04.7 8086 2021 1 ioatdma - - 00:00:56.553 + rm -f /tmp/spdk-ld-path 00:00:56.553 + source autorun-spdk.conf 00:00:56.553 ++ SPDK_RUN_FUNCTIONAL_TEST=1 00:00:56.553 ++ SPDK_TEST_NVMF=1 00:00:56.553 ++ SPDK_TEST_NVME_CLI=1 00:00:56.553 ++ SPDK_TEST_NVMF_TRANSPORT=tcp 00:00:56.553 ++ SPDK_TEST_NVMF_NICS=e810 00:00:56.553 ++ SPDK_TEST_VFIOUSER=1 00:00:56.553 ++ SPDK_RUN_UBSAN=1 00:00:56.553 ++ NET_TYPE=phy 00:00:56.553 ++ RUN_NIGHTLY=0 00:00:56.553 + (( SPDK_TEST_NVME_CMB == 1 || SPDK_TEST_NVME_PMR == 1 )) 00:00:56.553 + [[ -n '' ]] 00:00:56.553 + sudo git config --global --add safe.directory /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:56.553 + for M in /var/spdk/build-*-manifest.txt 00:00:56.553 + [[ -f /var/spdk/build-pkg-manifest.txt ]] 00:00:56.553 + cp /var/spdk/build-pkg-manifest.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/output/ 00:00:56.553 + for M in /var/spdk/build-*-manifest.txt 00:00:56.553 + [[ -f /var/spdk/build-repo-manifest.txt ]] 00:00:56.554 + cp /var/spdk/build-repo-manifest.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/output/ 00:00:56.554 ++ uname 00:00:56.554 + [[ Linux == \L\i\n\u\x ]] 00:00:56.554 + sudo dmesg -T 00:00:56.554 + sudo dmesg --clear 00:00:56.554 + dmesg_pid=298560 00:00:56.554 + [[ Fedora Linux == FreeBSD ]] 00:00:56.554 + export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:00:56.554 + UNBIND_ENTIRE_IOMMU_GROUP=yes 00:00:56.554 + [[ -e /var/spdk/dependencies/vhost/spdk_test_image.qcow2 ]] 00:00:56.554 + [[ -x /usr/src/fio-static/fio ]] 00:00:56.554 + export FIO_BIN=/usr/src/fio-static/fio 00:00:56.554 + FIO_BIN=/usr/src/fio-static/fio 00:00:56.554 + sudo dmesg -Tw 00:00:56.554 + [[ '' == \/\v\a\r\/\j\e\n\k\i\n\s\/\w\o\r\k\s\p\a\c\e\/\n\v\m\f\-\t\c\p\-\p\h\y\-\a\u\t\o\t\e\s\t\/\q\e\m\u\_\v\f\i\o\/* ]] 00:00:56.554 + [[ ! -v VFIO_QEMU_BIN ]] 00:00:56.554 + [[ -e /usr/local/qemu/vfio-user-latest ]] 00:00:56.554 + export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:00:56.554 + VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:00:56.554 + [[ -e /usr/local/qemu/vanilla-latest ]] 00:00:56.554 + export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:00:56.554 + QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:00:56.554 + spdk/autorun.sh /var/jenkins/workspace/nvmf-tcp-phy-autotest/autorun-spdk.conf 00:00:56.554 Test configuration: 00:00:56.554 SPDK_RUN_FUNCTIONAL_TEST=1 00:00:56.554 SPDK_TEST_NVMF=1 00:00:56.554 SPDK_TEST_NVME_CLI=1 00:00:56.554 SPDK_TEST_NVMF_TRANSPORT=tcp 00:00:56.554 SPDK_TEST_NVMF_NICS=e810 00:00:56.554 SPDK_TEST_VFIOUSER=1 00:00:56.554 SPDK_RUN_UBSAN=1 00:00:56.554 NET_TYPE=phy 00:00:56.554 RUN_NIGHTLY=0 17:46:50 -- common/autobuild_common.sh@15 -- $ source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:00:56.554 17:46:50 -- scripts/common.sh@508 -- $ [[ -e /bin/wpdk_common.sh ]] 00:00:56.554 17:46:50 -- scripts/common.sh@516 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:00:56.554 17:46:50 -- scripts/common.sh@517 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:00:56.554 17:46:50 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:00:56.554 17:46:50 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:00:56.554 17:46:50 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:00:56.554 17:46:50 -- paths/export.sh@5 -- $ export PATH 00:00:56.554 17:46:50 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:00:56.554 17:46:50 -- common/autobuild_common.sh@443 -- $ out=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output 00:00:56.554 17:46:50 -- common/autobuild_common.sh@444 -- $ date +%s 00:00:56.554 17:46:50 -- common/autobuild_common.sh@444 -- $ mktemp -dt spdk_1721058410.XXXXXX 00:00:56.554 17:46:50 -- common/autobuild_common.sh@444 -- $ SPDK_WORKSPACE=/tmp/spdk_1721058410.da0wPF 00:00:56.554 17:46:50 -- common/autobuild_common.sh@446 -- $ [[ -n '' ]] 00:00:56.554 17:46:50 -- common/autobuild_common.sh@450 -- $ '[' -n '' ']' 00:00:56.554 17:46:50 -- common/autobuild_common.sh@453 -- $ scanbuild_exclude='--exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/' 00:00:56.554 17:46:50 -- common/autobuild_common.sh@457 -- $ scanbuild_exclude+=' --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/xnvme --exclude /tmp' 00:00:56.554 17:46:50 -- common/autobuild_common.sh@459 -- $ scanbuild='scan-build -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/scan-build-tmp --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/ --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/xnvme --exclude /tmp --status-bugs' 00:00:56.554 17:46:50 -- common/autobuild_common.sh@460 -- $ get_config_params 00:00:56.554 17:46:50 -- common/autotest_common.sh@396 -- $ xtrace_disable 00:00:56.554 17:46:50 -- common/autotest_common.sh@10 -- $ set +x 00:00:56.554 17:46:50 -- common/autobuild_common.sh@460 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-coverage --with-ublk --with-vfio-user' 00:00:56.554 17:46:50 -- common/autobuild_common.sh@462 -- $ start_monitor_resources 00:00:56.554 17:46:50 -- pm/common@17 -- $ local monitor 00:00:56.554 17:46:50 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:00:56.554 17:46:50 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:00:56.554 17:46:50 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:00:56.554 17:46:50 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:00:56.554 17:46:50 -- pm/common@25 -- $ sleep 1 00:00:56.554 17:46:50 -- pm/common@21 -- $ date +%s 00:00:56.554 17:46:50 -- pm/common@21 -- $ date +%s 00:00:56.554 17:46:50 -- pm/common@21 -- $ date +%s 00:00:56.554 17:46:50 -- pm/common@21 -- $ date +%s 00:00:56.554 17:46:50 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-temp -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1721058410 00:00:56.554 17:46:50 -- pm/common@21 -- $ sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-bmc-pm -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1721058410 00:00:56.554 17:46:50 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-load -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1721058410 00:00:56.554 17:46:50 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-vmstat -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autobuild.sh.1721058410 00:00:56.554 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1721058410_collect-cpu-temp.pm.log 00:00:56.814 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1721058410_collect-vmstat.pm.log 00:00:56.814 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1721058410_collect-cpu-load.pm.log 00:00:56.814 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autobuild.sh.1721058410_collect-bmc-pm.bmc.pm.log 00:00:57.752 17:46:51 -- common/autobuild_common.sh@463 -- $ trap stop_monitor_resources EXIT 00:00:57.752 17:46:51 -- spdk/autobuild.sh@11 -- $ SPDK_TEST_AUTOBUILD= 00:00:57.752 17:46:51 -- spdk/autobuild.sh@12 -- $ umask 022 00:00:57.752 17:46:51 -- spdk/autobuild.sh@13 -- $ cd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:00:57.752 17:46:51 -- spdk/autobuild.sh@16 -- $ date -u 00:00:57.752 Mon Jul 15 03:46:51 PM UTC 2024 00:00:57.752 17:46:51 -- spdk/autobuild.sh@17 -- $ git describe --tags 00:00:57.752 v24.09-pre-212-g2f3522da7 00:00:57.752 17:46:51 -- spdk/autobuild.sh@19 -- $ '[' 0 -eq 1 ']' 00:00:57.752 17:46:51 -- spdk/autobuild.sh@23 -- $ '[' 1 -eq 1 ']' 00:00:57.752 17:46:51 -- spdk/autobuild.sh@24 -- $ run_test ubsan echo 'using ubsan' 00:00:57.752 17:46:51 -- common/autotest_common.sh@1099 -- $ '[' 3 -le 1 ']' 00:00:57.752 17:46:51 -- common/autotest_common.sh@1105 -- $ xtrace_disable 00:00:57.752 17:46:51 -- common/autotest_common.sh@10 -- $ set +x 00:00:57.752 ************************************ 00:00:57.752 START TEST ubsan 00:00:57.752 ************************************ 00:00:57.752 17:46:51 ubsan -- common/autotest_common.sh@1123 -- $ echo 'using ubsan' 00:00:57.752 using ubsan 00:00:57.752 00:00:57.752 real 0m0.000s 00:00:57.752 user 0m0.000s 00:00:57.752 sys 0m0.000s 00:00:57.752 17:46:51 ubsan -- common/autotest_common.sh@1124 -- $ xtrace_disable 00:00:57.752 17:46:51 ubsan -- common/autotest_common.sh@10 -- $ set +x 00:00:57.752 ************************************ 00:00:57.752 END TEST ubsan 00:00:57.752 ************************************ 00:00:57.752 17:46:51 -- common/autotest_common.sh@1142 -- $ return 0 00:00:57.752 17:46:51 -- spdk/autobuild.sh@27 -- $ '[' -n '' ']' 00:00:57.752 17:46:51 -- spdk/autobuild.sh@31 -- $ case "$SPDK_TEST_AUTOBUILD" in 00:00:57.752 17:46:51 -- spdk/autobuild.sh@47 -- $ [[ 0 -eq 1 ]] 00:00:57.752 17:46:51 -- spdk/autobuild.sh@51 -- $ [[ 0 -eq 1 ]] 00:00:57.752 17:46:51 -- spdk/autobuild.sh@55 -- $ [[ -n '' ]] 00:00:57.752 17:46:51 -- spdk/autobuild.sh@57 -- $ [[ 0 -eq 1 ]] 00:00:57.752 17:46:51 -- spdk/autobuild.sh@59 -- $ [[ 0 -eq 1 ]] 00:00:57.752 17:46:51 -- spdk/autobuild.sh@62 -- $ [[ 0 -eq 1 ]] 00:00:57.752 17:46:51 -- spdk/autobuild.sh@67 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/configure --enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-coverage --with-ublk --with-vfio-user --with-shared 00:00:57.752 Using default SPDK env in /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk 00:00:57.752 Using default DPDK in /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build 00:00:58.321 Using 'verbs' RDMA provider 00:01:11.094 Configuring ISA-L (logfile: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/.spdk-isal.log)...done. 00:01:21.080 Configuring ISA-L-crypto (logfile: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/.spdk-isal-crypto.log)...done. 00:01:21.080 Creating mk/config.mk...done. 00:01:21.080 Creating mk/cc.flags.mk...done. 00:01:21.080 Type 'make' to build. 00:01:21.080 17:47:14 -- spdk/autobuild.sh@69 -- $ run_test make make -j96 00:01:21.080 17:47:14 -- common/autotest_common.sh@1099 -- $ '[' 3 -le 1 ']' 00:01:21.080 17:47:14 -- common/autotest_common.sh@1105 -- $ xtrace_disable 00:01:21.080 17:47:14 -- common/autotest_common.sh@10 -- $ set +x 00:01:21.080 ************************************ 00:01:21.081 START TEST make 00:01:21.081 ************************************ 00:01:21.081 17:47:14 make -- common/autotest_common.sh@1123 -- $ make -j96 00:01:21.340 make[1]: Nothing to be done for 'all'. 00:01:22.773 The Meson build system 00:01:22.773 Version: 1.3.1 00:01:22.773 Source dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/libvfio-user 00:01:22.773 Build dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug 00:01:22.773 Build type: native build 00:01:22.773 Project name: libvfio-user 00:01:22.773 Project version: 0.0.1 00:01:22.773 C compiler for the host machine: cc (gcc 13.2.1 "cc (GCC) 13.2.1 20231011 (Red Hat 13.2.1-4)") 00:01:22.773 C linker for the host machine: cc ld.bfd 2.39-16 00:01:22.773 Host machine cpu family: x86_64 00:01:22.773 Host machine cpu: x86_64 00:01:22.773 Run-time dependency threads found: YES 00:01:22.773 Library dl found: YES 00:01:22.773 Found pkg-config: YES (/usr/bin/pkg-config) 1.8.0 00:01:22.773 Run-time dependency json-c found: YES 0.17 00:01:22.773 Run-time dependency cmocka found: YES 1.1.7 00:01:22.773 Program pytest-3 found: NO 00:01:22.773 Program flake8 found: NO 00:01:22.773 Program misspell-fixer found: NO 00:01:22.773 Program restructuredtext-lint found: NO 00:01:22.773 Program valgrind found: YES (/usr/bin/valgrind) 00:01:22.773 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:01:22.773 Compiler for C supports arguments -Wmissing-declarations: YES 00:01:22.773 Compiler for C supports arguments -Wwrite-strings: YES 00:01:22.773 ../libvfio-user/test/meson.build:20: WARNING: Project targets '>= 0.53.0' but uses feature introduced in '0.57.0': exclude_suites arg in add_test_setup. 00:01:22.773 Program test-lspci.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/libvfio-user/test/test-lspci.sh) 00:01:22.773 Program test-linkage.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/libvfio-user/test/test-linkage.sh) 00:01:22.773 ../libvfio-user/test/py/meson.build:16: WARNING: Project targets '>= 0.53.0' but uses feature introduced in '0.57.0': exclude_suites arg in add_test_setup. 00:01:22.773 Build targets in project: 8 00:01:22.773 WARNING: Project specifies a minimum meson_version '>= 0.53.0' but uses features which were added in newer versions: 00:01:22.773 * 0.57.0: {'exclude_suites arg in add_test_setup'} 00:01:22.773 00:01:22.773 libvfio-user 0.0.1 00:01:22.773 00:01:22.773 User defined options 00:01:22.773 buildtype : debug 00:01:22.773 default_library: shared 00:01:22.773 libdir : /usr/local/lib 00:01:22.773 00:01:22.773 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:01:23.031 ninja: Entering directory `/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug' 00:01:23.290 [1/37] Compiling C object samples/null.p/null.c.o 00:01:23.290 [2/37] Compiling C object samples/lspci.p/lspci.c.o 00:01:23.290 [3/37] Compiling C object samples/shadow_ioeventfd_server.p/shadow_ioeventfd_server.c.o 00:01:23.290 [4/37] Compiling C object lib/libvfio-user.so.0.0.1.p/tran.c.o 00:01:23.290 [5/37] Compiling C object samples/client.p/.._lib_tran.c.o 00:01:23.290 [6/37] Compiling C object samples/gpio-pci-idio-16.p/gpio-pci-idio-16.c.o 00:01:23.290 [7/37] Compiling C object test/unit_tests.p/.._lib_irq.c.o 00:01:23.290 [8/37] Compiling C object lib/libvfio-user.so.0.0.1.p/pci.c.o 00:01:23.290 [9/37] Compiling C object test/unit_tests.p/.._lib_tran.c.o 00:01:23.290 [10/37] Compiling C object test/unit_tests.p/.._lib_migration.c.o 00:01:23.290 [11/37] Compiling C object lib/libvfio-user.so.0.0.1.p/irq.c.o 00:01:23.290 [12/37] Compiling C object test/unit_tests.p/.._lib_tran_pipe.c.o 00:01:23.290 [13/37] Compiling C object samples/client.p/.._lib_migration.c.o 00:01:23.290 [14/37] Compiling C object lib/libvfio-user.so.0.0.1.p/migration.c.o 00:01:23.290 [15/37] Compiling C object test/unit_tests.p/.._lib_pci.c.o 00:01:23.290 [16/37] Compiling C object test/unit_tests.p/mocks.c.o 00:01:23.290 [17/37] Compiling C object test/unit_tests.p/.._lib_dma.c.o 00:01:23.290 [18/37] Compiling C object test/unit_tests.p/unit-tests.c.o 00:01:23.290 [19/37] Compiling C object samples/client.p/.._lib_tran_sock.c.o 00:01:23.290 [20/37] Compiling C object test/unit_tests.p/.._lib_tran_sock.c.o 00:01:23.290 [21/37] Compiling C object lib/libvfio-user.so.0.0.1.p/pci_caps.c.o 00:01:23.290 [22/37] Compiling C object lib/libvfio-user.so.0.0.1.p/dma.c.o 00:01:23.290 [23/37] Compiling C object lib/libvfio-user.so.0.0.1.p/tran_sock.c.o 00:01:23.290 [24/37] Compiling C object test/unit_tests.p/.._lib_pci_caps.c.o 00:01:23.290 [25/37] Compiling C object samples/server.p/server.c.o 00:01:23.290 [26/37] Compiling C object samples/client.p/client.c.o 00:01:23.290 [27/37] Linking target samples/client 00:01:23.290 [28/37] Compiling C object lib/libvfio-user.so.0.0.1.p/libvfio-user.c.o 00:01:23.548 [29/37] Compiling C object test/unit_tests.p/.._lib_libvfio-user.c.o 00:01:23.548 [30/37] Linking target lib/libvfio-user.so.0.0.1 00:01:23.548 [31/37] Linking target test/unit_tests 00:01:23.548 [32/37] Generating symbol file lib/libvfio-user.so.0.0.1.p/libvfio-user.so.0.0.1.symbols 00:01:23.548 [33/37] Linking target samples/server 00:01:23.548 [34/37] Linking target samples/gpio-pci-idio-16 00:01:23.548 [35/37] Linking target samples/null 00:01:23.548 [36/37] Linking target samples/lspci 00:01:23.548 [37/37] Linking target samples/shadow_ioeventfd_server 00:01:23.548 INFO: autodetecting backend as ninja 00:01:23.548 INFO: calculating backend command to run: /usr/local/bin/ninja -C /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug 00:01:23.548 DESTDIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user meson install --quiet -C /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug 00:01:24.113 ninja: Entering directory `/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/build-debug' 00:01:24.113 ninja: no work to do. 00:01:29.379 The Meson build system 00:01:29.379 Version: 1.3.1 00:01:29.379 Source dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk 00:01:29.379 Build dir: /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build-tmp 00:01:29.379 Build type: native build 00:01:29.379 Program cat found: YES (/usr/bin/cat) 00:01:29.379 Project name: DPDK 00:01:29.379 Project version: 24.03.0 00:01:29.379 C compiler for the host machine: cc (gcc 13.2.1 "cc (GCC) 13.2.1 20231011 (Red Hat 13.2.1-4)") 00:01:29.379 C linker for the host machine: cc ld.bfd 2.39-16 00:01:29.379 Host machine cpu family: x86_64 00:01:29.379 Host machine cpu: x86_64 00:01:29.379 Message: ## Building in Developer Mode ## 00:01:29.379 Program pkg-config found: YES (/usr/bin/pkg-config) 00:01:29.379 Program check-symbols.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/buildtools/check-symbols.sh) 00:01:29.379 Program options-ibverbs-static.sh found: YES (/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/buildtools/options-ibverbs-static.sh) 00:01:29.379 Program python3 found: YES (/usr/bin/python3) 00:01:29.379 Program cat found: YES (/usr/bin/cat) 00:01:29.379 Compiler for C supports arguments -march=native: YES 00:01:29.379 Checking for size of "void *" : 8 00:01:29.379 Checking for size of "void *" : 8 (cached) 00:01:29.379 Compiler for C supports link arguments -Wl,--undefined-version: NO 00:01:29.379 Library m found: YES 00:01:29.379 Library numa found: YES 00:01:29.379 Has header "numaif.h" : YES 00:01:29.379 Library fdt found: NO 00:01:29.379 Library execinfo found: NO 00:01:29.379 Has header "execinfo.h" : YES 00:01:29.379 Found pkg-config: YES (/usr/bin/pkg-config) 1.8.0 00:01:29.379 Run-time dependency libarchive found: NO (tried pkgconfig) 00:01:29.379 Run-time dependency libbsd found: NO (tried pkgconfig) 00:01:29.379 Run-time dependency jansson found: NO (tried pkgconfig) 00:01:29.379 Run-time dependency openssl found: YES 3.0.9 00:01:29.379 Run-time dependency libpcap found: YES 1.10.4 00:01:29.379 Has header "pcap.h" with dependency libpcap: YES 00:01:29.379 Compiler for C supports arguments -Wcast-qual: YES 00:01:29.379 Compiler for C supports arguments -Wdeprecated: YES 00:01:29.379 Compiler for C supports arguments -Wformat: YES 00:01:29.379 Compiler for C supports arguments -Wformat-nonliteral: NO 00:01:29.379 Compiler for C supports arguments -Wformat-security: NO 00:01:29.379 Compiler for C supports arguments -Wmissing-declarations: YES 00:01:29.380 Compiler for C supports arguments -Wmissing-prototypes: YES 00:01:29.380 Compiler for C supports arguments -Wnested-externs: YES 00:01:29.380 Compiler for C supports arguments -Wold-style-definition: YES 00:01:29.380 Compiler for C supports arguments -Wpointer-arith: YES 00:01:29.380 Compiler for C supports arguments -Wsign-compare: YES 00:01:29.380 Compiler for C supports arguments -Wstrict-prototypes: YES 00:01:29.380 Compiler for C supports arguments -Wundef: YES 00:01:29.380 Compiler for C supports arguments -Wwrite-strings: YES 00:01:29.380 Compiler for C supports arguments -Wno-address-of-packed-member: YES 00:01:29.380 Compiler for C supports arguments -Wno-packed-not-aligned: YES 00:01:29.380 Compiler for C supports arguments -Wno-missing-field-initializers: YES 00:01:29.380 Compiler for C supports arguments -Wno-zero-length-bounds: YES 00:01:29.380 Program objdump found: YES (/usr/bin/objdump) 00:01:29.380 Compiler for C supports arguments -mavx512f: YES 00:01:29.380 Checking if "AVX512 checking" compiles: YES 00:01:29.380 Fetching value of define "__SSE4_2__" : 1 00:01:29.380 Fetching value of define "__AES__" : 1 00:01:29.380 Fetching value of define "__AVX__" : 1 00:01:29.380 Fetching value of define "__AVX2__" : 1 00:01:29.380 Fetching value of define "__AVX512BW__" : 1 00:01:29.380 Fetching value of define "__AVX512CD__" : 1 00:01:29.380 Fetching value of define "__AVX512DQ__" : 1 00:01:29.380 Fetching value of define "__AVX512F__" : 1 00:01:29.380 Fetching value of define "__AVX512VL__" : 1 00:01:29.380 Fetching value of define "__PCLMUL__" : 1 00:01:29.380 Fetching value of define "__RDRND__" : 1 00:01:29.380 Fetching value of define "__RDSEED__" : 1 00:01:29.380 Fetching value of define "__VPCLMULQDQ__" : (undefined) 00:01:29.380 Fetching value of define "__znver1__" : (undefined) 00:01:29.380 Fetching value of define "__znver2__" : (undefined) 00:01:29.380 Fetching value of define "__znver3__" : (undefined) 00:01:29.380 Fetching value of define "__znver4__" : (undefined) 00:01:29.380 Compiler for C supports arguments -Wno-format-truncation: YES 00:01:29.380 Message: lib/log: Defining dependency "log" 00:01:29.380 Message: lib/kvargs: Defining dependency "kvargs" 00:01:29.380 Message: lib/telemetry: Defining dependency "telemetry" 00:01:29.380 Checking for function "getentropy" : NO 00:01:29.380 Message: lib/eal: Defining dependency "eal" 00:01:29.380 Message: lib/ring: Defining dependency "ring" 00:01:29.380 Message: lib/rcu: Defining dependency "rcu" 00:01:29.380 Message: lib/mempool: Defining dependency "mempool" 00:01:29.380 Message: lib/mbuf: Defining dependency "mbuf" 00:01:29.380 Fetching value of define "__PCLMUL__" : 1 (cached) 00:01:29.380 Fetching value of define "__AVX512F__" : 1 (cached) 00:01:29.380 Fetching value of define "__AVX512BW__" : 1 (cached) 00:01:29.380 Fetching value of define "__AVX512DQ__" : 1 (cached) 00:01:29.380 Fetching value of define "__AVX512VL__" : 1 (cached) 00:01:29.380 Fetching value of define "__VPCLMULQDQ__" : (undefined) (cached) 00:01:29.380 Compiler for C supports arguments -mpclmul: YES 00:01:29.380 Compiler for C supports arguments -maes: YES 00:01:29.380 Compiler for C supports arguments -mavx512f: YES (cached) 00:01:29.380 Compiler for C supports arguments -mavx512bw: YES 00:01:29.380 Compiler for C supports arguments -mavx512dq: YES 00:01:29.380 Compiler for C supports arguments -mavx512vl: YES 00:01:29.380 Compiler for C supports arguments -mvpclmulqdq: YES 00:01:29.380 Compiler for C supports arguments -mavx2: YES 00:01:29.380 Compiler for C supports arguments -mavx: YES 00:01:29.380 Message: lib/net: Defining dependency "net" 00:01:29.380 Message: lib/meter: Defining dependency "meter" 00:01:29.380 Message: lib/ethdev: Defining dependency "ethdev" 00:01:29.380 Message: lib/pci: Defining dependency "pci" 00:01:29.380 Message: lib/cmdline: Defining dependency "cmdline" 00:01:29.380 Message: lib/hash: Defining dependency "hash" 00:01:29.380 Message: lib/timer: Defining dependency "timer" 00:01:29.380 Message: lib/compressdev: Defining dependency "compressdev" 00:01:29.380 Message: lib/cryptodev: Defining dependency "cryptodev" 00:01:29.380 Message: lib/dmadev: Defining dependency "dmadev" 00:01:29.380 Compiler for C supports arguments -Wno-cast-qual: YES 00:01:29.380 Message: lib/power: Defining dependency "power" 00:01:29.380 Message: lib/reorder: Defining dependency "reorder" 00:01:29.380 Message: lib/security: Defining dependency "security" 00:01:29.380 Has header "linux/userfaultfd.h" : YES 00:01:29.380 Has header "linux/vduse.h" : YES 00:01:29.380 Message: lib/vhost: Defining dependency "vhost" 00:01:29.380 Compiler for C supports arguments -Wno-format-truncation: YES (cached) 00:01:29.380 Message: drivers/bus/pci: Defining dependency "bus_pci" 00:01:29.380 Message: drivers/bus/vdev: Defining dependency "bus_vdev" 00:01:29.380 Message: drivers/mempool/ring: Defining dependency "mempool_ring" 00:01:29.380 Message: Disabling raw/* drivers: missing internal dependency "rawdev" 00:01:29.380 Message: Disabling regex/* drivers: missing internal dependency "regexdev" 00:01:29.380 Message: Disabling ml/* drivers: missing internal dependency "mldev" 00:01:29.380 Message: Disabling event/* drivers: missing internal dependency "eventdev" 00:01:29.380 Message: Disabling baseband/* drivers: missing internal dependency "bbdev" 00:01:29.380 Message: Disabling gpu/* drivers: missing internal dependency "gpudev" 00:01:29.380 Program doxygen found: YES (/usr/bin/doxygen) 00:01:29.380 Configuring doxy-api-html.conf using configuration 00:01:29.380 Configuring doxy-api-man.conf using configuration 00:01:29.380 Program mandb found: YES (/usr/bin/mandb) 00:01:29.380 Program sphinx-build found: NO 00:01:29.380 Configuring rte_build_config.h using configuration 00:01:29.380 Message: 00:01:29.380 ================= 00:01:29.380 Applications Enabled 00:01:29.380 ================= 00:01:29.380 00:01:29.380 apps: 00:01:29.380 00:01:29.380 00:01:29.380 Message: 00:01:29.380 ================= 00:01:29.380 Libraries Enabled 00:01:29.380 ================= 00:01:29.380 00:01:29.380 libs: 00:01:29.380 log, kvargs, telemetry, eal, ring, rcu, mempool, mbuf, 00:01:29.380 net, meter, ethdev, pci, cmdline, hash, timer, compressdev, 00:01:29.380 cryptodev, dmadev, power, reorder, security, vhost, 00:01:29.380 00:01:29.380 Message: 00:01:29.380 =============== 00:01:29.380 Drivers Enabled 00:01:29.380 =============== 00:01:29.380 00:01:29.380 common: 00:01:29.380 00:01:29.380 bus: 00:01:29.380 pci, vdev, 00:01:29.380 mempool: 00:01:29.380 ring, 00:01:29.380 dma: 00:01:29.380 00:01:29.380 net: 00:01:29.380 00:01:29.380 crypto: 00:01:29.380 00:01:29.380 compress: 00:01:29.380 00:01:29.380 vdpa: 00:01:29.380 00:01:29.380 00:01:29.380 Message: 00:01:29.380 ================= 00:01:29.380 Content Skipped 00:01:29.380 ================= 00:01:29.380 00:01:29.380 apps: 00:01:29.380 dumpcap: explicitly disabled via build config 00:01:29.380 graph: explicitly disabled via build config 00:01:29.380 pdump: explicitly disabled via build config 00:01:29.380 proc-info: explicitly disabled via build config 00:01:29.380 test-acl: explicitly disabled via build config 00:01:29.380 test-bbdev: explicitly disabled via build config 00:01:29.380 test-cmdline: explicitly disabled via build config 00:01:29.380 test-compress-perf: explicitly disabled via build config 00:01:29.380 test-crypto-perf: explicitly disabled via build config 00:01:29.380 test-dma-perf: explicitly disabled via build config 00:01:29.380 test-eventdev: explicitly disabled via build config 00:01:29.380 test-fib: explicitly disabled via build config 00:01:29.380 test-flow-perf: explicitly disabled via build config 00:01:29.380 test-gpudev: explicitly disabled via build config 00:01:29.380 test-mldev: explicitly disabled via build config 00:01:29.380 test-pipeline: explicitly disabled via build config 00:01:29.380 test-pmd: explicitly disabled via build config 00:01:29.380 test-regex: explicitly disabled via build config 00:01:29.380 test-sad: explicitly disabled via build config 00:01:29.380 test-security-perf: explicitly disabled via build config 00:01:29.380 00:01:29.380 libs: 00:01:29.380 argparse: explicitly disabled via build config 00:01:29.380 metrics: explicitly disabled via build config 00:01:29.380 acl: explicitly disabled via build config 00:01:29.380 bbdev: explicitly disabled via build config 00:01:29.380 bitratestats: explicitly disabled via build config 00:01:29.380 bpf: explicitly disabled via build config 00:01:29.380 cfgfile: explicitly disabled via build config 00:01:29.380 distributor: explicitly disabled via build config 00:01:29.380 efd: explicitly disabled via build config 00:01:29.380 eventdev: explicitly disabled via build config 00:01:29.380 dispatcher: explicitly disabled via build config 00:01:29.380 gpudev: explicitly disabled via build config 00:01:29.380 gro: explicitly disabled via build config 00:01:29.380 gso: explicitly disabled via build config 00:01:29.380 ip_frag: explicitly disabled via build config 00:01:29.380 jobstats: explicitly disabled via build config 00:01:29.380 latencystats: explicitly disabled via build config 00:01:29.380 lpm: explicitly disabled via build config 00:01:29.380 member: explicitly disabled via build config 00:01:29.380 pcapng: explicitly disabled via build config 00:01:29.380 rawdev: explicitly disabled via build config 00:01:29.380 regexdev: explicitly disabled via build config 00:01:29.380 mldev: explicitly disabled via build config 00:01:29.380 rib: explicitly disabled via build config 00:01:29.380 sched: explicitly disabled via build config 00:01:29.380 stack: explicitly disabled via build config 00:01:29.380 ipsec: explicitly disabled via build config 00:01:29.380 pdcp: explicitly disabled via build config 00:01:29.380 fib: explicitly disabled via build config 00:01:29.380 port: explicitly disabled via build config 00:01:29.380 pdump: explicitly disabled via build config 00:01:29.380 table: explicitly disabled via build config 00:01:29.380 pipeline: explicitly disabled via build config 00:01:29.380 graph: explicitly disabled via build config 00:01:29.380 node: explicitly disabled via build config 00:01:29.380 00:01:29.380 drivers: 00:01:29.380 common/cpt: not in enabled drivers build config 00:01:29.380 common/dpaax: not in enabled drivers build config 00:01:29.380 common/iavf: not in enabled drivers build config 00:01:29.380 common/idpf: not in enabled drivers build config 00:01:29.380 common/ionic: not in enabled drivers build config 00:01:29.380 common/mvep: not in enabled drivers build config 00:01:29.380 common/octeontx: not in enabled drivers build config 00:01:29.380 bus/auxiliary: not in enabled drivers build config 00:01:29.380 bus/cdx: not in enabled drivers build config 00:01:29.380 bus/dpaa: not in enabled drivers build config 00:01:29.380 bus/fslmc: not in enabled drivers build config 00:01:29.380 bus/ifpga: not in enabled drivers build config 00:01:29.380 bus/platform: not in enabled drivers build config 00:01:29.380 bus/uacce: not in enabled drivers build config 00:01:29.380 bus/vmbus: not in enabled drivers build config 00:01:29.380 common/cnxk: not in enabled drivers build config 00:01:29.381 common/mlx5: not in enabled drivers build config 00:01:29.381 common/nfp: not in enabled drivers build config 00:01:29.381 common/nitrox: not in enabled drivers build config 00:01:29.381 common/qat: not in enabled drivers build config 00:01:29.381 common/sfc_efx: not in enabled drivers build config 00:01:29.381 mempool/bucket: not in enabled drivers build config 00:01:29.381 mempool/cnxk: not in enabled drivers build config 00:01:29.381 mempool/dpaa: not in enabled drivers build config 00:01:29.381 mempool/dpaa2: not in enabled drivers build config 00:01:29.381 mempool/octeontx: not in enabled drivers build config 00:01:29.381 mempool/stack: not in enabled drivers build config 00:01:29.381 dma/cnxk: not in enabled drivers build config 00:01:29.381 dma/dpaa: not in enabled drivers build config 00:01:29.381 dma/dpaa2: not in enabled drivers build config 00:01:29.381 dma/hisilicon: not in enabled drivers build config 00:01:29.381 dma/idxd: not in enabled drivers build config 00:01:29.381 dma/ioat: not in enabled drivers build config 00:01:29.381 dma/skeleton: not in enabled drivers build config 00:01:29.381 net/af_packet: not in enabled drivers build config 00:01:29.381 net/af_xdp: not in enabled drivers build config 00:01:29.381 net/ark: not in enabled drivers build config 00:01:29.381 net/atlantic: not in enabled drivers build config 00:01:29.381 net/avp: not in enabled drivers build config 00:01:29.381 net/axgbe: not in enabled drivers build config 00:01:29.381 net/bnx2x: not in enabled drivers build config 00:01:29.381 net/bnxt: not in enabled drivers build config 00:01:29.381 net/bonding: not in enabled drivers build config 00:01:29.381 net/cnxk: not in enabled drivers build config 00:01:29.381 net/cpfl: not in enabled drivers build config 00:01:29.381 net/cxgbe: not in enabled drivers build config 00:01:29.381 net/dpaa: not in enabled drivers build config 00:01:29.381 net/dpaa2: not in enabled drivers build config 00:01:29.381 net/e1000: not in enabled drivers build config 00:01:29.381 net/ena: not in enabled drivers build config 00:01:29.381 net/enetc: not in enabled drivers build config 00:01:29.381 net/enetfec: not in enabled drivers build config 00:01:29.381 net/enic: not in enabled drivers build config 00:01:29.381 net/failsafe: not in enabled drivers build config 00:01:29.381 net/fm10k: not in enabled drivers build config 00:01:29.381 net/gve: not in enabled drivers build config 00:01:29.381 net/hinic: not in enabled drivers build config 00:01:29.381 net/hns3: not in enabled drivers build config 00:01:29.381 net/i40e: not in enabled drivers build config 00:01:29.381 net/iavf: not in enabled drivers build config 00:01:29.381 net/ice: not in enabled drivers build config 00:01:29.381 net/idpf: not in enabled drivers build config 00:01:29.381 net/igc: not in enabled drivers build config 00:01:29.381 net/ionic: not in enabled drivers build config 00:01:29.381 net/ipn3ke: not in enabled drivers build config 00:01:29.381 net/ixgbe: not in enabled drivers build config 00:01:29.381 net/mana: not in enabled drivers build config 00:01:29.381 net/memif: not in enabled drivers build config 00:01:29.381 net/mlx4: not in enabled drivers build config 00:01:29.381 net/mlx5: not in enabled drivers build config 00:01:29.381 net/mvneta: not in enabled drivers build config 00:01:29.381 net/mvpp2: not in enabled drivers build config 00:01:29.381 net/netvsc: not in enabled drivers build config 00:01:29.381 net/nfb: not in enabled drivers build config 00:01:29.381 net/nfp: not in enabled drivers build config 00:01:29.381 net/ngbe: not in enabled drivers build config 00:01:29.381 net/null: not in enabled drivers build config 00:01:29.381 net/octeontx: not in enabled drivers build config 00:01:29.381 net/octeon_ep: not in enabled drivers build config 00:01:29.381 net/pcap: not in enabled drivers build config 00:01:29.381 net/pfe: not in enabled drivers build config 00:01:29.381 net/qede: not in enabled drivers build config 00:01:29.381 net/ring: not in enabled drivers build config 00:01:29.381 net/sfc: not in enabled drivers build config 00:01:29.381 net/softnic: not in enabled drivers build config 00:01:29.381 net/tap: not in enabled drivers build config 00:01:29.381 net/thunderx: not in enabled drivers build config 00:01:29.381 net/txgbe: not in enabled drivers build config 00:01:29.381 net/vdev_netvsc: not in enabled drivers build config 00:01:29.381 net/vhost: not in enabled drivers build config 00:01:29.381 net/virtio: not in enabled drivers build config 00:01:29.381 net/vmxnet3: not in enabled drivers build config 00:01:29.381 raw/*: missing internal dependency, "rawdev" 00:01:29.381 crypto/armv8: not in enabled drivers build config 00:01:29.381 crypto/bcmfs: not in enabled drivers build config 00:01:29.381 crypto/caam_jr: not in enabled drivers build config 00:01:29.381 crypto/ccp: not in enabled drivers build config 00:01:29.381 crypto/cnxk: not in enabled drivers build config 00:01:29.381 crypto/dpaa_sec: not in enabled drivers build config 00:01:29.381 crypto/dpaa2_sec: not in enabled drivers build config 00:01:29.381 crypto/ipsec_mb: not in enabled drivers build config 00:01:29.381 crypto/mlx5: not in enabled drivers build config 00:01:29.381 crypto/mvsam: not in enabled drivers build config 00:01:29.381 crypto/nitrox: not in enabled drivers build config 00:01:29.381 crypto/null: not in enabled drivers build config 00:01:29.381 crypto/octeontx: not in enabled drivers build config 00:01:29.381 crypto/openssl: not in enabled drivers build config 00:01:29.381 crypto/scheduler: not in enabled drivers build config 00:01:29.381 crypto/uadk: not in enabled drivers build config 00:01:29.381 crypto/virtio: not in enabled drivers build config 00:01:29.381 compress/isal: not in enabled drivers build config 00:01:29.381 compress/mlx5: not in enabled drivers build config 00:01:29.381 compress/nitrox: not in enabled drivers build config 00:01:29.381 compress/octeontx: not in enabled drivers build config 00:01:29.381 compress/zlib: not in enabled drivers build config 00:01:29.381 regex/*: missing internal dependency, "regexdev" 00:01:29.381 ml/*: missing internal dependency, "mldev" 00:01:29.381 vdpa/ifc: not in enabled drivers build config 00:01:29.381 vdpa/mlx5: not in enabled drivers build config 00:01:29.381 vdpa/nfp: not in enabled drivers build config 00:01:29.381 vdpa/sfc: not in enabled drivers build config 00:01:29.381 event/*: missing internal dependency, "eventdev" 00:01:29.381 baseband/*: missing internal dependency, "bbdev" 00:01:29.381 gpu/*: missing internal dependency, "gpudev" 00:01:29.381 00:01:29.381 00:01:29.381 Build targets in project: 85 00:01:29.381 00:01:29.381 DPDK 24.03.0 00:01:29.381 00:01:29.381 User defined options 00:01:29.381 buildtype : debug 00:01:29.381 default_library : shared 00:01:29.381 libdir : lib 00:01:29.381 prefix : /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build 00:01:29.381 c_args : -Wno-stringop-overflow -fcommon -Wno-stringop-overread -Wno-array-bounds -fPIC -Werror 00:01:29.381 c_link_args : 00:01:29.381 cpu_instruction_set: native 00:01:29.381 disable_apps : test-dma-perf,test,test-sad,test-acl,test-pmd,test-mldev,test-compress-perf,test-cmdline,test-regex,test-fib,graph,test-bbdev,dumpcap,test-gpudev,proc-info,test-pipeline,test-flow-perf,test-crypto-perf,pdump,test-eventdev,test-security-perf 00:01:29.381 disable_libs : port,lpm,ipsec,regexdev,dispatcher,argparse,bitratestats,rawdev,stack,graph,acl,bbdev,pipeline,member,sched,pcapng,mldev,eventdev,efd,metrics,latencystats,cfgfile,ip_frag,jobstats,pdump,pdcp,rib,node,fib,distributor,gso,table,bpf,gpudev,gro 00:01:29.381 enable_docs : false 00:01:29.381 enable_drivers : bus,bus/pci,bus/vdev,mempool/ring 00:01:29.381 enable_kmods : false 00:01:29.381 max_lcores : 128 00:01:29.381 tests : false 00:01:29.381 00:01:29.381 Found ninja-1.11.1.git.kitware.jobserver-1 at /usr/local/bin/ninja 00:01:29.381 ninja: Entering directory `/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build-tmp' 00:01:29.381 [1/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hypervisor.c.o 00:01:29.381 [2/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_errno.c.o 00:01:29.381 [3/268] Compiling C object lib/librte_kvargs.a.p/kvargs_rte_kvargs.c.o 00:01:29.381 [4/268] Linking static target lib/librte_kvargs.a 00:01:29.381 [5/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_cpuflags.c.o 00:01:29.381 [6/268] Compiling C object lib/librte_eal.a.p/eal_x86_rte_hypervisor.c.o 00:01:29.381 [7/268] Compiling C object lib/librte_eal.a.p/eal_common_rte_version.c.o 00:01:29.381 [8/268] Compiling C object lib/librte_log.a.p/log_log_linux.c.o 00:01:29.381 [9/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_hexdump.c.o 00:01:29.381 [10/268] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cpuflags.c.o 00:01:29.381 [11/268] Compiling C object lib/librte_eal.a.p/eal_common_rte_reciprocal.c.o 00:01:29.381 [12/268] Compiling C object lib/librte_eal.a.p/eal_unix_eal_debug.c.o 00:01:29.381 [13/268] Compiling C object lib/librte_eal.a.p/eal_x86_rte_spinlock.c.o 00:01:29.381 [14/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_uuid.c.o 00:01:29.381 [15/268] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_data.c.o 00:01:29.381 [16/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_string_fns.c.o 00:01:29.381 [17/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_cpuflags.c.o 00:01:29.381 [18/268] Compiling C object lib/librte_log.a.p/log_log.c.o 00:01:29.381 [19/268] Linking static target lib/librte_log.a 00:01:29.381 [20/268] Compiling C object lib/librte_pci.a.p/pci_rte_pci.c.o 00:01:29.381 [21/268] Linking static target lib/librte_pci.a 00:01:29.381 [22/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline.c.o 00:01:29.381 [23/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse.c.o 00:01:29.646 [24/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_cirbuf.c.o 00:01:29.646 [25/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_num.c.o 00:01:29.646 [26/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_debug.c.o 00:01:29.646 [27/268] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_timer.c.o 00:01:29.646 [28/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_class.c.o 00:01:29.646 [29/268] Compiling C object lib/librte_eal.a.p/eal_common_rte_random.c.o 00:01:29.646 [30/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_timer.c.o 00:01:29.646 [31/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_portlist.c.o 00:01:29.646 [32/268] Compiling C object lib/librte_eal.a.p/eal_common_rte_keepalive.c.o 00:01:29.646 [33/268] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_thread.c.o 00:01:29.646 [34/268] Compiling C object lib/librte_eal.a.p/eal_unix_rte_thread.c.o 00:01:29.646 [35/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_bus.c.o 00:01:29.646 [36/268] Compiling C object lib/librte_eal.a.p/eal_unix_eal_file.c.o 00:01:29.646 [37/268] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry_legacy.c.o 00:01:29.646 [38/268] Compiling C object lib/librte_net.a.p/net_rte_net_crc.c.o 00:01:29.646 [39/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio_mp_sync.c.o 00:01:29.646 [40/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_thread.c.o 00:01:29.646 [41/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_ctf.c.o 00:01:29.646 [42/268] Compiling C object lib/librte_meter.a.p/meter_rte_meter.c.o 00:01:29.646 [43/268] Compiling C object lib/librte_eal.a.p/eal_common_hotplug_mp.c.o 00:01:29.646 [44/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_vt100.c.o 00:01:29.646 [45/268] Compiling C object lib/librte_eal.a.p/eal_common_malloc_mp.c.o 00:01:29.646 [46/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_string.c.o 00:01:29.646 [47/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_timer.c.o 00:01:29.646 [48/268] Linking static target lib/librte_meter.a 00:01:29.646 [49/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace.c.o 00:01:29.646 [50/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_config.c.o 00:01:29.646 [51/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_alarm.c.o 00:01:29.646 [52/268] Compiling C object lib/net/libnet_crc_avx512_lib.a.p/net_crc_avx512.c.o 00:01:29.646 [53/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_tailqs.c.o 00:01:29.646 [54/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_thread.c.o 00:01:29.646 [55/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_utils.c.o 00:01:29.646 [56/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_socket.c.o 00:01:29.646 [57/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_devargs.c.o 00:01:29.646 [58/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_mcfg.c.o 00:01:29.646 [59/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memalloc.c.o 00:01:29.646 [60/268] Compiling C object lib/librte_net.a.p/net_net_crc_sse.c.o 00:01:29.646 [61/268] Generating lib/kvargs.sym_chk with a custom command (wrapped by meson to capture output) 00:01:29.646 [62/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_os_unix.c.o 00:01:29.646 [63/268] Compiling C object lib/librte_hash.a.p/hash_rte_hash_crc.c.o 00:01:29.646 [64/268] Linking static target lib/net/libnet_crc_avx512_lib.a 00:01:29.646 [65/268] Compiling C object lib/librte_eal.a.p/eal_x86_rte_cycles.c.o 00:01:29.646 [66/268] Compiling C object lib/librte_eal.a.p/eal_x86_rte_power_intrinsics.c.o 00:01:29.646 [67/268] Compiling C object lib/librte_mempool.a.p/mempool_mempool_trace_points.c.o 00:01:29.646 [68/268] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_ptype.c.o 00:01:29.646 [69/268] Compiling C object lib/librte_telemetry.a.p/telemetry_telemetry.c.o 00:01:29.646 [70/268] Compiling C object lib/librte_eal.a.p/eal_unix_eal_firmware.c.o 00:01:29.646 [71/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_lcore.c.o 00:01:29.646 [72/268] Compiling C object lib/librte_ring.a.p/ring_rte_ring.c.o 00:01:29.646 [73/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_launch.c.o 00:01:29.646 [74/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dynmem.c.o 00:01:29.646 [75/268] Compiling C object lib/librte_eal.a.p/eal_unix_eal_filesystem.c.o 00:01:29.646 [76/268] Compiling C object lib/librte_eal.a.p/eal_unix_eal_unix_memory.c.o 00:01:29.646 [77/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_trace_points.c.o 00:01:29.646 [78/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_interrupts.c.o 00:01:29.646 [79/268] Compiling C object lib/librte_eal.a.p/eal_common_malloc_elem.c.o 00:01:29.646 [80/268] Compiling C object lib/librte_eal.a.p/eal_common_malloc_heap.c.o 00:01:29.646 [81/268] Compiling C object lib/librte_eal.a.p/eal_common_rte_malloc.c.o 00:01:29.646 [82/268] Linking static target lib/librte_telemetry.a 00:01:29.646 [83/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_dev.c.o 00:01:29.646 [84/268] Linking static target lib/librte_ring.a 00:01:29.908 [85/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memzone.c.o 00:01:29.908 [86/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_lcore.c.o 00:01:29.908 [87/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_memory.c.o 00:01:29.908 [88/268] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_pool_ops.c.o 00:01:29.908 [89/268] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops_default.c.o 00:01:29.908 [90/268] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool_ops.c.o 00:01:29.908 [91/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_dev.c.o 00:01:29.908 [92/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_proc.c.o 00:01:29.908 [93/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_rdline.c.o 00:01:29.908 [94/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_hugepage_info.c.o 00:01:29.908 [95/268] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_profile.c.o 00:01:29.908 [96/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_interrupts.c.o 00:01:29.908 [97/268] Compiling C object lib/librte_net.a.p/net_rte_ether.c.o 00:01:29.908 [98/268] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf_dyn.c.o 00:01:29.908 [99/268] Compiling C object lib/librte_power.a.p/power_power_kvm_vm.c.o 00:01:29.908 [100/268] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_class_eth.c.o 00:01:29.908 [101/268] Compiling C object lib/librte_eal.a.p/eal_common_rte_service.c.o 00:01:29.908 [102/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_fbarray.c.o 00:01:29.908 [103/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memalloc.c.o 00:01:29.908 [104/268] Compiling C object lib/librte_power.a.p/power_power_common.c.o 00:01:29.908 [105/268] Compiling C object lib/librte_net.a.p/net_rte_net.c.o 00:01:29.908 [106/268] Generating lib/pci.sym_chk with a custom command (wrapped by meson to capture output) 00:01:29.908 [107/268] Compiling C object lib/librte_power.a.p/power_guest_channel.c.o 00:01:29.908 [108/268] Compiling C object lib/librte_rcu.a.p/rcu_rte_rcu_qsbr.c.o 00:01:29.908 [109/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_vfio.c.o 00:01:29.908 [110/268] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_private.c.o 00:01:29.908 [111/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal.c.o 00:01:29.908 [112/268] Linking static target lib/librte_rcu.a 00:01:29.908 [113/268] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev_params.c.o 00:01:29.908 [114/268] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8472.c.o 00:01:29.908 [115/268] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_params.c.o 00:01:29.908 [116/268] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_telemetry.c.o 00:01:29.908 [117/268] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_linux_ethtool.c.o 00:01:29.908 [118/268] Compiling C object lib/librte_eal.a.p/eal_linux_eal_memory.c.o 00:01:29.908 [119/268] Compiling C object lib/librte_mempool.a.p/mempool_rte_mempool.c.o 00:01:29.908 [120/268] Compiling C object lib/librte_net.a.p/net_rte_arp.c.o 00:01:29.908 [121/268] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_telemetry.c.o 00:01:29.908 [122/268] Compiling C object lib/librte_vhost.a.p/vhost_fd_man.c.o 00:01:29.908 [123/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_etheraddr.c.o 00:01:29.908 [124/268] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_driver.c.o 00:01:29.908 [125/268] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8079.c.o 00:01:29.908 [126/268] Linking static target lib/librte_mempool.a 00:01:29.908 [127/268] Linking static target lib/librte_net.a 00:01:29.908 [128/268] Compiling C object lib/librte_eal.a.p/eal_common_eal_common_options.c.o 00:01:29.908 [129/268] Linking static target lib/librte_eal.a 00:01:29.908 [130/268] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev_cman.c.o 00:01:29.908 [131/268] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_8636.c.o 00:01:29.908 [132/268] Compiling C object lib/librte_cmdline.a.p/cmdline_cmdline_parse_ipaddr.c.o 00:01:29.908 [133/268] Linking static target lib/librte_cmdline.a 00:01:29.908 [134/268] Compiling C object lib/librte_ethdev.a.p/ethdev_sff_common.c.o 00:01:29.908 [135/268] Generating lib/meter.sym_chk with a custom command (wrapped by meson to capture output) 00:01:30.166 [136/268] Generating lib/ring.sym_chk with a custom command (wrapped by meson to capture output) 00:01:30.166 [137/268] Generating lib/log.sym_chk with a custom command (wrapped by meson to capture output) 00:01:30.166 [138/268] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_mtr.c.o 00:01:30.166 [139/268] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev_pmd.c.o 00:01:30.166 [140/268] Compiling C object lib/librte_mbuf.a.p/mbuf_rte_mbuf.c.o 00:01:30.166 [141/268] Linking target lib/librte_log.so.24.1 00:01:30.166 [142/268] Linking static target lib/librte_mbuf.a 00:01:30.166 [143/268] Generating lib/rcu.sym_chk with a custom command (wrapped by meson to capture output) 00:01:30.166 [144/268] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_compressdev.c.o 00:01:30.166 [145/268] Compiling C object lib/librte_power.a.p/power_power_intel_uncore.c.o 00:01:30.166 [146/268] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev_trace_points.c.o 00:01:30.166 [147/268] Compiling C object lib/librte_power.a.p/power_power_amd_pstate_cpufreq.c.o 00:01:30.166 [148/268] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common_uio.c.o 00:01:30.166 [149/268] Compiling C object lib/librte_hash.a.p/hash_rte_thash_gfni.c.o 00:01:30.166 [150/268] Generating lib/telemetry.sym_chk with a custom command (wrapped by meson to capture output) 00:01:30.166 [151/268] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_tm.c.o 00:01:30.166 [152/268] Generating lib/net.sym_chk with a custom command (wrapped by meson to capture output) 00:01:30.166 [153/268] Compiling C object lib/librte_timer.a.p/timer_rte_timer.c.o 00:01:30.166 [154/268] Compiling C object lib/librte_hash.a.p/hash_rte_fbk_hash.c.o 00:01:30.166 [155/268] Linking static target lib/librte_timer.a 00:01:30.166 [156/268] Compiling C object lib/librte_hash.a.p/hash_rte_thash.c.o 00:01:30.166 [157/268] Compiling C object lib/librte_power.a.p/power_rte_power.c.o 00:01:30.166 [158/268] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_trace_points.c.o 00:01:30.166 [159/268] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_uio.c.o 00:01:30.166 [160/268] Compiling C object lib/librte_vhost.a.p/vhost_vdpa.c.o 00:01:30.166 [161/268] Compiling C object lib/librte_power.a.p/power_rte_power_uncore.c.o 00:01:30.166 [162/268] Compiling C object lib/librte_power.a.p/power_power_acpi_cpufreq.c.o 00:01:30.166 [163/268] Compiling C object lib/librte_ethdev.a.p/ethdev_ethdev_trace_points.c.o 00:01:30.166 [164/268] Generating symbol file lib/librte_log.so.24.1.p/librte_log.so.24.1.symbols 00:01:30.166 [165/268] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net_ctrl.c.o 00:01:30.166 [166/268] Compiling C object lib/librte_power.a.p/power_power_cppc_cpufreq.c.o 00:01:30.166 [167/268] Compiling C object lib/librte_compressdev.a.p/compressdev_rte_comp.c.o 00:01:30.166 [168/268] Compiling C object lib/librte_reorder.a.p/reorder_rte_reorder.c.o 00:01:30.166 [169/268] Linking static target lib/librte_compressdev.a 00:01:30.166 [170/268] Linking static target lib/librte_reorder.a 00:01:30.166 [171/268] Linking target lib/librte_kvargs.so.24.1 00:01:30.166 [172/268] Compiling C object lib/librte_vhost.a.p/vhost_socket.c.o 00:01:30.166 [173/268] Linking target lib/librte_telemetry.so.24.1 00:01:30.166 [174/268] Compiling C object lib/librte_vhost.a.p/vhost_iotlb.c.o 00:01:30.166 [175/268] Compiling C object lib/librte_cryptodev.a.p/cryptodev_cryptodev_pmd.c.o 00:01:30.166 [176/268] Compiling C object drivers/libtmp_rte_mempool_ring.a.p/mempool_ring_rte_mempool_ring.c.o 00:01:30.425 [177/268] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci_vfio.c.o 00:01:30.425 [178/268] Linking static target drivers/libtmp_rte_mempool_ring.a 00:01:30.425 [179/268] Compiling C object lib/librte_power.a.p/power_power_pstate_cpufreq.c.o 00:01:30.425 [180/268] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_flow.c.o 00:01:30.425 [181/268] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_linux_pci.c.o 00:01:30.425 [182/268] Compiling C object drivers/libtmp_rte_bus_pci.a.p/bus_pci_pci_common.c.o 00:01:30.425 [183/268] Compiling C object lib/librte_dmadev.a.p/dmadev_rte_dmadev.c.o 00:01:30.425 [184/268] Compiling C object drivers/libtmp_rte_bus_vdev.a.p/bus_vdev_vdev.c.o 00:01:30.425 [185/268] Linking static target lib/librte_dmadev.a 00:01:30.425 [186/268] Linking static target drivers/libtmp_rte_bus_vdev.a 00:01:30.425 [187/268] Compiling C object lib/librte_security.a.p/security_rte_security.c.o 00:01:30.425 [188/268] Linking static target drivers/libtmp_rte_bus_pci.a 00:01:30.425 [189/268] Linking static target lib/librte_security.a 00:01:30.425 [190/268] Compiling C object lib/librte_vhost.a.p/vhost_vduse.c.o 00:01:30.425 [191/268] Generating symbol file lib/librte_telemetry.so.24.1.p/librte_telemetry.so.24.1.symbols 00:01:30.425 [192/268] Compiling C object lib/librte_power.a.p/power_rte_power_pmd_mgmt.c.o 00:01:30.425 [193/268] Generating symbol file lib/librte_kvargs.so.24.1.p/librte_kvargs.so.24.1.symbols 00:01:30.425 [194/268] Linking static target lib/librte_power.a 00:01:30.425 [195/268] Compiling C object lib/librte_vhost.a.p/vhost_vhost_user.c.o 00:01:30.425 [196/268] Compiling C object lib/librte_vhost.a.p/vhost_vhost.c.o 00:01:30.425 [197/268] Generating drivers/rte_mempool_ring.pmd.c with a custom command 00:01:30.425 [198/268] Compiling C object lib/librte_hash.a.p/hash_rte_cuckoo_hash.c.o 00:01:30.425 [199/268] Compiling C object drivers/librte_mempool_ring.so.24.1.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:01:30.425 [200/268] Compiling C object drivers/librte_mempool_ring.a.p/meson-generated_.._rte_mempool_ring.pmd.c.o 00:01:30.425 [201/268] Linking static target lib/librte_hash.a 00:01:30.425 [202/268] Linking static target drivers/librte_mempool_ring.a 00:01:30.425 [203/268] Generating drivers/rte_bus_vdev.pmd.c with a custom command 00:01:30.425 [204/268] Generating drivers/rte_bus_pci.pmd.c with a custom command 00:01:30.684 [205/268] Compiling C object drivers/librte_bus_vdev.so.24.1.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:01:30.684 [206/268] Compiling C object drivers/librte_bus_pci.a.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:01:30.684 [207/268] Compiling C object drivers/librte_bus_vdev.a.p/meson-generated_.._rte_bus_vdev.pmd.c.o 00:01:30.684 [208/268] Compiling C object drivers/librte_bus_pci.so.24.1.p/meson-generated_.._rte_bus_pci.pmd.c.o 00:01:30.684 [209/268] Linking static target drivers/librte_bus_pci.a 00:01:30.684 [210/268] Linking static target drivers/librte_bus_vdev.a 00:01:30.684 [211/268] Generating lib/mempool.sym_chk with a custom command (wrapped by meson to capture output) 00:01:30.684 [212/268] Generating lib/timer.sym_chk with a custom command (wrapped by meson to capture output) 00:01:30.684 [213/268] Generating lib/reorder.sym_chk with a custom command (wrapped by meson to capture output) 00:01:30.684 [214/268] Compiling C object lib/librte_cryptodev.a.p/cryptodev_rte_cryptodev.c.o 00:01:30.684 [215/268] Linking static target lib/librte_cryptodev.a 00:01:30.684 [216/268] Generating lib/mbuf.sym_chk with a custom command (wrapped by meson to capture output) 00:01:30.942 [217/268] Compiling C object lib/librte_ethdev.a.p/ethdev_rte_ethdev.c.o 00:01:30.942 [218/268] Linking static target lib/librte_ethdev.a 00:01:30.942 [219/268] Generating lib/compressdev.sym_chk with a custom command (wrapped by meson to capture output) 00:01:30.942 [220/268] Generating drivers/rte_bus_vdev.sym_chk with a custom command (wrapped by meson to capture output) 00:01:30.942 [221/268] Generating lib/security.sym_chk with a custom command (wrapped by meson to capture output) 00:01:30.942 [222/268] Generating lib/dmadev.sym_chk with a custom command (wrapped by meson to capture output) 00:01:30.942 [223/268] Generating lib/cmdline.sym_chk with a custom command (wrapped by meson to capture output) 00:01:30.942 [224/268] Compiling C object lib/librte_vhost.a.p/vhost_vhost_crypto.c.o 00:01:31.199 [225/268] Generating lib/power.sym_chk with a custom command (wrapped by meson to capture output) 00:01:31.199 [226/268] Generating drivers/rte_bus_pci.sym_chk with a custom command (wrapped by meson to capture output) 00:01:31.457 [227/268] Generating lib/hash.sym_chk with a custom command (wrapped by meson to capture output) 00:01:32.022 [228/268] Compiling C object lib/librte_vhost.a.p/vhost_virtio_net.c.o 00:01:32.022 [229/268] Linking static target lib/librte_vhost.a 00:01:32.622 [230/268] Generating lib/cryptodev.sym_chk with a custom command (wrapped by meson to capture output) 00:01:33.994 [231/268] Generating lib/vhost.sym_chk with a custom command (wrapped by meson to capture output) 00:01:39.250 [232/268] Generating lib/ethdev.sym_chk with a custom command (wrapped by meson to capture output) 00:01:39.507 [233/268] Generating lib/eal.sym_chk with a custom command (wrapped by meson to capture output) 00:01:39.507 [234/268] Linking target lib/librte_eal.so.24.1 00:01:39.507 [235/268] Generating symbol file lib/librte_eal.so.24.1.p/librte_eal.so.24.1.symbols 00:01:39.765 [236/268] Linking target lib/librte_ring.so.24.1 00:01:39.765 [237/268] Linking target lib/librte_pci.so.24.1 00:01:39.765 [238/268] Linking target lib/librte_dmadev.so.24.1 00:01:39.765 [239/268] Linking target lib/librte_timer.so.24.1 00:01:39.765 [240/268] Linking target lib/librte_meter.so.24.1 00:01:39.765 [241/268] Linking target drivers/librte_bus_vdev.so.24.1 00:01:39.765 [242/268] Generating symbol file lib/librte_meter.so.24.1.p/librte_meter.so.24.1.symbols 00:01:39.765 [243/268] Generating symbol file lib/librte_dmadev.so.24.1.p/librte_dmadev.so.24.1.symbols 00:01:39.765 [244/268] Generating symbol file lib/librte_pci.so.24.1.p/librte_pci.so.24.1.symbols 00:01:39.765 [245/268] Generating symbol file lib/librte_ring.so.24.1.p/librte_ring.so.24.1.symbols 00:01:39.765 [246/268] Generating symbol file lib/librte_timer.so.24.1.p/librte_timer.so.24.1.symbols 00:01:39.765 [247/268] Linking target lib/librte_rcu.so.24.1 00:01:39.765 [248/268] Linking target lib/librte_mempool.so.24.1 00:01:39.765 [249/268] Linking target drivers/librte_bus_pci.so.24.1 00:01:40.022 [250/268] Generating symbol file lib/librte_rcu.so.24.1.p/librte_rcu.so.24.1.symbols 00:01:40.022 [251/268] Generating symbol file lib/librte_mempool.so.24.1.p/librte_mempool.so.24.1.symbols 00:01:40.022 [252/268] Linking target lib/librte_mbuf.so.24.1 00:01:40.022 [253/268] Linking target drivers/librte_mempool_ring.so.24.1 00:01:40.279 [254/268] Generating symbol file lib/librte_mbuf.so.24.1.p/librte_mbuf.so.24.1.symbols 00:01:40.279 [255/268] Linking target lib/librte_reorder.so.24.1 00:01:40.279 [256/268] Linking target lib/librte_compressdev.so.24.1 00:01:40.279 [257/268] Linking target lib/librte_net.so.24.1 00:01:40.279 [258/268] Linking target lib/librte_cryptodev.so.24.1 00:01:40.279 [259/268] Generating symbol file lib/librte_net.so.24.1.p/librte_net.so.24.1.symbols 00:01:40.279 [260/268] Generating symbol file lib/librte_cryptodev.so.24.1.p/librte_cryptodev.so.24.1.symbols 00:01:40.279 [261/268] Linking target lib/librte_cmdline.so.24.1 00:01:40.279 [262/268] Linking target lib/librte_security.so.24.1 00:01:40.279 [263/268] Linking target lib/librte_hash.so.24.1 00:01:40.536 [264/268] Linking target lib/librte_ethdev.so.24.1 00:01:40.536 [265/268] Generating symbol file lib/librte_hash.so.24.1.p/librte_hash.so.24.1.symbols 00:01:40.536 [266/268] Generating symbol file lib/librte_ethdev.so.24.1.p/librte_ethdev.so.24.1.symbols 00:01:40.536 [267/268] Linking target lib/librte_power.so.24.1 00:01:40.536 [268/268] Linking target lib/librte_vhost.so.24.1 00:01:40.536 INFO: autodetecting backend as ninja 00:01:40.537 INFO: calculating backend command to run: /usr/local/bin/ninja -C /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build-tmp -j 96 00:01:41.466 CC lib/ut_mock/mock.o 00:01:41.466 CC lib/log/log.o 00:01:41.466 CC lib/log/log_flags.o 00:01:41.466 CC lib/log/log_deprecated.o 00:01:41.466 CC lib/ut/ut.o 00:01:41.723 LIB libspdk_ut_mock.a 00:01:41.723 LIB libspdk_log.a 00:01:41.723 LIB libspdk_ut.a 00:01:41.723 SO libspdk_ut_mock.so.6.0 00:01:41.723 SO libspdk_log.so.7.0 00:01:41.723 SO libspdk_ut.so.2.0 00:01:41.723 SYMLINK libspdk_ut_mock.so 00:01:41.723 SYMLINK libspdk_ut.so 00:01:41.723 SYMLINK libspdk_log.so 00:01:41.980 CC lib/util/base64.o 00:01:41.980 CC lib/util/cpuset.o 00:01:41.980 CC lib/util/crc16.o 00:01:41.980 CC lib/util/bit_array.o 00:01:41.980 CC lib/util/crc32c.o 00:01:41.980 CC lib/util/crc32.o 00:01:41.980 CC lib/util/crc32_ieee.o 00:01:41.980 CC lib/util/crc64.o 00:01:41.980 CC lib/util/dif.o 00:01:41.980 CC lib/util/hexlify.o 00:01:41.980 CC lib/util/fd.o 00:01:41.980 CC lib/util/file.o 00:01:41.980 CC lib/util/pipe.o 00:01:41.980 CC lib/util/iov.o 00:01:41.980 CC lib/util/math.o 00:01:41.980 CC lib/util/strerror_tls.o 00:01:41.980 CC lib/util/uuid.o 00:01:41.980 CC lib/util/string.o 00:01:41.980 CC lib/util/fd_group.o 00:01:41.980 CC lib/util/xor.o 00:01:41.980 CC lib/util/zipf.o 00:01:41.980 CC lib/dma/dma.o 00:01:41.980 CXX lib/trace_parser/trace.o 00:01:41.980 CC lib/ioat/ioat.o 00:01:42.238 CC lib/vfio_user/host/vfio_user_pci.o 00:01:42.238 CC lib/vfio_user/host/vfio_user.o 00:01:42.238 LIB libspdk_dma.a 00:01:42.238 SO libspdk_dma.so.4.0 00:01:42.238 LIB libspdk_ioat.a 00:01:42.238 SYMLINK libspdk_dma.so 00:01:42.531 SO libspdk_ioat.so.7.0 00:01:42.531 LIB libspdk_vfio_user.a 00:01:42.531 SYMLINK libspdk_ioat.so 00:01:42.531 SO libspdk_vfio_user.so.5.0 00:01:42.531 LIB libspdk_util.a 00:01:42.531 SYMLINK libspdk_vfio_user.so 00:01:42.531 SO libspdk_util.so.9.1 00:01:42.795 SYMLINK libspdk_util.so 00:01:42.795 LIB libspdk_trace_parser.a 00:01:42.795 SO libspdk_trace_parser.so.5.0 00:01:42.795 SYMLINK libspdk_trace_parser.so 00:01:43.053 CC lib/conf/conf.o 00:01:43.053 CC lib/json/json_parse.o 00:01:43.053 CC lib/json/json_util.o 00:01:43.053 CC lib/json/json_write.o 00:01:43.053 CC lib/env_dpdk/env.o 00:01:43.053 CC lib/env_dpdk/memory.o 00:01:43.053 CC lib/env_dpdk/init.o 00:01:43.053 CC lib/env_dpdk/pci.o 00:01:43.053 CC lib/env_dpdk/threads.o 00:01:43.053 CC lib/env_dpdk/pci_ioat.o 00:01:43.053 CC lib/env_dpdk/pci_vmd.o 00:01:43.053 CC lib/env_dpdk/pci_virtio.o 00:01:43.053 CC lib/vmd/vmd.o 00:01:43.053 CC lib/env_dpdk/pci_idxd.o 00:01:43.053 CC lib/vmd/led.o 00:01:43.053 CC lib/env_dpdk/sigbus_handler.o 00:01:43.053 CC lib/env_dpdk/pci_event.o 00:01:43.053 CC lib/env_dpdk/pci_dpdk_2207.o 00:01:43.053 CC lib/env_dpdk/pci_dpdk.o 00:01:43.053 CC lib/env_dpdk/pci_dpdk_2211.o 00:01:43.053 CC lib/idxd/idxd.o 00:01:43.053 CC lib/rdma_provider/common.o 00:01:43.053 CC lib/rdma_utils/rdma_utils.o 00:01:43.053 CC lib/rdma_provider/rdma_provider_verbs.o 00:01:43.053 CC lib/idxd/idxd_user.o 00:01:43.053 CC lib/idxd/idxd_kernel.o 00:01:43.053 LIB libspdk_rdma_provider.a 00:01:43.311 LIB libspdk_conf.a 00:01:43.311 SO libspdk_rdma_provider.so.6.0 00:01:43.311 LIB libspdk_json.a 00:01:43.311 SO libspdk_conf.so.6.0 00:01:43.311 LIB libspdk_rdma_utils.a 00:01:43.311 SO libspdk_json.so.6.0 00:01:43.311 SYMLINK libspdk_rdma_provider.so 00:01:43.311 SO libspdk_rdma_utils.so.1.0 00:01:43.311 SYMLINK libspdk_conf.so 00:01:43.311 SYMLINK libspdk_json.so 00:01:43.311 SYMLINK libspdk_rdma_utils.so 00:01:43.311 LIB libspdk_idxd.a 00:01:43.568 LIB libspdk_vmd.a 00:01:43.569 SO libspdk_idxd.so.12.0 00:01:43.569 SO libspdk_vmd.so.6.0 00:01:43.569 SYMLINK libspdk_idxd.so 00:01:43.569 SYMLINK libspdk_vmd.so 00:01:43.569 CC lib/jsonrpc/jsonrpc_server.o 00:01:43.569 CC lib/jsonrpc/jsonrpc_server_tcp.o 00:01:43.569 CC lib/jsonrpc/jsonrpc_client.o 00:01:43.569 CC lib/jsonrpc/jsonrpc_client_tcp.o 00:01:43.826 LIB libspdk_jsonrpc.a 00:01:43.826 SO libspdk_jsonrpc.so.6.0 00:01:43.826 SYMLINK libspdk_jsonrpc.so 00:01:44.085 LIB libspdk_env_dpdk.a 00:01:44.085 SO libspdk_env_dpdk.so.14.1 00:01:44.085 SYMLINK libspdk_env_dpdk.so 00:01:44.343 CC lib/rpc/rpc.o 00:01:44.343 LIB libspdk_rpc.a 00:01:44.343 SO libspdk_rpc.so.6.0 00:01:44.602 SYMLINK libspdk_rpc.so 00:01:44.859 CC lib/notify/notify.o 00:01:44.859 CC lib/notify/notify_rpc.o 00:01:44.859 CC lib/trace/trace.o 00:01:44.859 CC lib/trace/trace_flags.o 00:01:44.859 CC lib/trace/trace_rpc.o 00:01:44.859 CC lib/keyring/keyring.o 00:01:44.859 CC lib/keyring/keyring_rpc.o 00:01:44.859 LIB libspdk_notify.a 00:01:44.859 SO libspdk_notify.so.6.0 00:01:45.117 SYMLINK libspdk_notify.so 00:01:45.117 LIB libspdk_trace.a 00:01:45.117 LIB libspdk_keyring.a 00:01:45.117 SO libspdk_trace.so.10.0 00:01:45.117 SO libspdk_keyring.so.1.0 00:01:45.117 SYMLINK libspdk_trace.so 00:01:45.117 SYMLINK libspdk_keyring.so 00:01:45.374 CC lib/sock/sock.o 00:01:45.374 CC lib/sock/sock_rpc.o 00:01:45.374 CC lib/thread/thread.o 00:01:45.374 CC lib/thread/iobuf.o 00:01:45.631 LIB libspdk_sock.a 00:01:45.631 SO libspdk_sock.so.10.0 00:01:45.631 SYMLINK libspdk_sock.so 00:01:46.194 CC lib/nvme/nvme_ctrlr_cmd.o 00:01:46.194 CC lib/nvme/nvme_ctrlr.o 00:01:46.194 CC lib/nvme/nvme_fabric.o 00:01:46.194 CC lib/nvme/nvme_ns_cmd.o 00:01:46.194 CC lib/nvme/nvme_ns.o 00:01:46.194 CC lib/nvme/nvme_pcie_common.o 00:01:46.194 CC lib/nvme/nvme_pcie.o 00:01:46.194 CC lib/nvme/nvme_qpair.o 00:01:46.194 CC lib/nvme/nvme_quirks.o 00:01:46.194 CC lib/nvme/nvme.o 00:01:46.194 CC lib/nvme/nvme_transport.o 00:01:46.194 CC lib/nvme/nvme_discovery.o 00:01:46.194 CC lib/nvme/nvme_ctrlr_ocssd_cmd.o 00:01:46.194 CC lib/nvme/nvme_ns_ocssd_cmd.o 00:01:46.194 CC lib/nvme/nvme_tcp.o 00:01:46.194 CC lib/nvme/nvme_poll_group.o 00:01:46.194 CC lib/nvme/nvme_opal.o 00:01:46.194 CC lib/nvme/nvme_io_msg.o 00:01:46.194 CC lib/nvme/nvme_zns.o 00:01:46.194 CC lib/nvme/nvme_stubs.o 00:01:46.194 CC lib/nvme/nvme_auth.o 00:01:46.194 CC lib/nvme/nvme_cuse.o 00:01:46.194 CC lib/nvme/nvme_vfio_user.o 00:01:46.194 CC lib/nvme/nvme_rdma.o 00:01:46.451 LIB libspdk_thread.a 00:01:46.451 SO libspdk_thread.so.10.1 00:01:46.451 SYMLINK libspdk_thread.so 00:01:46.709 CC lib/blob/blobstore.o 00:01:46.709 CC lib/blob/zeroes.o 00:01:46.709 CC lib/blob/request.o 00:01:46.709 CC lib/blob/blob_bs_dev.o 00:01:46.709 CC lib/init/json_config.o 00:01:46.709 CC lib/init/subsystem.o 00:01:46.709 CC lib/init/subsystem_rpc.o 00:01:46.709 CC lib/init/rpc.o 00:01:46.709 CC lib/accel/accel_rpc.o 00:01:46.709 CC lib/accel/accel.o 00:01:46.709 CC lib/accel/accel_sw.o 00:01:46.709 CC lib/vfu_tgt/tgt_endpoint.o 00:01:46.709 CC lib/vfu_tgt/tgt_rpc.o 00:01:46.709 CC lib/virtio/virtio.o 00:01:46.709 CC lib/virtio/virtio_vhost_user.o 00:01:46.709 CC lib/virtio/virtio_pci.o 00:01:46.709 CC lib/virtio/virtio_vfio_user.o 00:01:46.967 LIB libspdk_init.a 00:01:46.967 SO libspdk_init.so.5.0 00:01:46.967 LIB libspdk_vfu_tgt.a 00:01:46.967 LIB libspdk_virtio.a 00:01:46.967 SYMLINK libspdk_init.so 00:01:47.224 SO libspdk_vfu_tgt.so.3.0 00:01:47.224 SO libspdk_virtio.so.7.0 00:01:47.224 SYMLINK libspdk_vfu_tgt.so 00:01:47.224 SYMLINK libspdk_virtio.so 00:01:47.481 CC lib/event/app.o 00:01:47.481 CC lib/event/log_rpc.o 00:01:47.481 CC lib/event/reactor.o 00:01:47.481 CC lib/event/app_rpc.o 00:01:47.481 CC lib/event/scheduler_static.o 00:01:47.481 LIB libspdk_accel.a 00:01:47.481 SO libspdk_accel.so.15.1 00:01:47.739 SYMLINK libspdk_accel.so 00:01:47.739 LIB libspdk_nvme.a 00:01:47.739 LIB libspdk_event.a 00:01:47.739 SO libspdk_nvme.so.13.1 00:01:47.739 SO libspdk_event.so.14.0 00:01:47.739 SYMLINK libspdk_event.so 00:01:47.996 CC lib/bdev/bdev.o 00:01:47.996 CC lib/bdev/bdev_rpc.o 00:01:47.996 CC lib/bdev/bdev_zone.o 00:01:47.996 CC lib/bdev/part.o 00:01:47.997 CC lib/bdev/scsi_nvme.o 00:01:47.997 SYMLINK libspdk_nvme.so 00:01:48.929 LIB libspdk_blob.a 00:01:48.929 SO libspdk_blob.so.11.0 00:01:48.929 SYMLINK libspdk_blob.so 00:01:49.186 CC lib/blobfs/blobfs.o 00:01:49.186 CC lib/blobfs/tree.o 00:01:49.186 CC lib/lvol/lvol.o 00:01:49.751 LIB libspdk_bdev.a 00:01:49.751 SO libspdk_bdev.so.15.1 00:01:49.751 SYMLINK libspdk_bdev.so 00:01:49.751 LIB libspdk_blobfs.a 00:01:49.751 SO libspdk_blobfs.so.10.0 00:01:50.009 LIB libspdk_lvol.a 00:01:50.009 SYMLINK libspdk_blobfs.so 00:01:50.009 SO libspdk_lvol.so.10.0 00:01:50.009 SYMLINK libspdk_lvol.so 00:01:50.009 CC lib/nbd/nbd.o 00:01:50.009 CC lib/nbd/nbd_rpc.o 00:01:50.009 CC lib/nvmf/ctrlr.o 00:01:50.009 CC lib/scsi/dev.o 00:01:50.009 CC lib/nvmf/ctrlr_discovery.o 00:01:50.009 CC lib/scsi/lun.o 00:01:50.009 CC lib/nvmf/ctrlr_bdev.o 00:01:50.009 CC lib/scsi/port.o 00:01:50.009 CC lib/scsi/scsi.o 00:01:50.009 CC lib/nvmf/nvmf_rpc.o 00:01:50.009 CC lib/nvmf/subsystem.o 00:01:50.009 CC lib/scsi/scsi_bdev.o 00:01:50.009 CC lib/nvmf/nvmf.o 00:01:50.009 CC lib/nvmf/transport.o 00:01:50.009 CC lib/scsi/scsi_pr.o 00:01:50.009 CC lib/scsi/scsi_rpc.o 00:01:50.009 CC lib/ublk/ublk.o 00:01:50.009 CC lib/scsi/task.o 00:01:50.009 CC lib/nvmf/tcp.o 00:01:50.009 CC lib/ublk/ublk_rpc.o 00:01:50.009 CC lib/nvmf/stubs.o 00:01:50.009 CC lib/nvmf/mdns_server.o 00:01:50.009 CC lib/nvmf/vfio_user.o 00:01:50.009 CC lib/nvmf/rdma.o 00:01:50.009 CC lib/nvmf/auth.o 00:01:50.009 CC lib/ftl/ftl_init.o 00:01:50.009 CC lib/ftl/ftl_core.o 00:01:50.009 CC lib/ftl/ftl_layout.o 00:01:50.009 CC lib/ftl/ftl_debug.o 00:01:50.009 CC lib/ftl/ftl_io.o 00:01:50.009 CC lib/ftl/ftl_sb.o 00:01:50.009 CC lib/ftl/ftl_l2p.o 00:01:50.009 CC lib/ftl/ftl_l2p_flat.o 00:01:50.009 CC lib/ftl/ftl_nv_cache.o 00:01:50.009 CC lib/ftl/ftl_band.o 00:01:50.009 CC lib/ftl/ftl_band_ops.o 00:01:50.009 CC lib/ftl/ftl_writer.o 00:01:50.009 CC lib/ftl/ftl_rq.o 00:01:50.009 CC lib/ftl/ftl_reloc.o 00:01:50.009 CC lib/ftl/ftl_l2p_cache.o 00:01:50.009 CC lib/ftl/ftl_p2l.o 00:01:50.009 CC lib/ftl/mngt/ftl_mngt.o 00:01:50.009 CC lib/ftl/mngt/ftl_mngt_bdev.o 00:01:50.009 CC lib/ftl/mngt/ftl_mngt_shutdown.o 00:01:50.009 CC lib/ftl/mngt/ftl_mngt_startup.o 00:01:50.009 CC lib/ftl/mngt/ftl_mngt_misc.o 00:01:50.009 CC lib/ftl/mngt/ftl_mngt_md.o 00:01:50.009 CC lib/ftl/mngt/ftl_mngt_ioch.o 00:01:50.009 CC lib/ftl/mngt/ftl_mngt_l2p.o 00:01:50.009 CC lib/ftl/mngt/ftl_mngt_band.o 00:01:50.009 CC lib/ftl/mngt/ftl_mngt_p2l.o 00:01:50.009 CC lib/ftl/mngt/ftl_mngt_upgrade.o 00:01:50.009 CC lib/ftl/mngt/ftl_mngt_self_test.o 00:01:50.009 CC lib/ftl/mngt/ftl_mngt_recovery.o 00:01:50.009 CC lib/ftl/utils/ftl_md.o 00:01:50.267 CC lib/ftl/utils/ftl_conf.o 00:01:50.267 CC lib/ftl/utils/ftl_mempool.o 00:01:50.267 CC lib/ftl/utils/ftl_bitmap.o 00:01:50.267 CC lib/ftl/utils/ftl_property.o 00:01:50.267 CC lib/ftl/utils/ftl_layout_tracker_bdev.o 00:01:50.267 CC lib/ftl/upgrade/ftl_sb_upgrade.o 00:01:50.267 CC lib/ftl/upgrade/ftl_p2l_upgrade.o 00:01:50.267 CC lib/ftl/upgrade/ftl_band_upgrade.o 00:01:50.267 CC lib/ftl/upgrade/ftl_layout_upgrade.o 00:01:50.267 CC lib/ftl/upgrade/ftl_chunk_upgrade.o 00:01:50.267 CC lib/ftl/upgrade/ftl_sb_v3.o 00:01:50.267 CC lib/ftl/upgrade/ftl_sb_v5.o 00:01:50.268 CC lib/ftl/upgrade/ftl_trim_upgrade.o 00:01:50.268 CC lib/ftl/nvc/ftl_nvc_bdev_vss.o 00:01:50.268 CC lib/ftl/base/ftl_base_dev.o 00:01:50.268 CC lib/ftl/base/ftl_base_bdev.o 00:01:50.268 CC lib/ftl/nvc/ftl_nvc_dev.o 00:01:50.268 CC lib/ftl/ftl_trace.o 00:01:50.525 LIB libspdk_nbd.a 00:01:50.525 SO libspdk_nbd.so.7.0 00:01:50.784 SYMLINK libspdk_nbd.so 00:01:50.784 LIB libspdk_scsi.a 00:01:50.784 LIB libspdk_ublk.a 00:01:50.784 SO libspdk_ublk.so.3.0 00:01:50.784 SO libspdk_scsi.so.9.0 00:01:50.784 SYMLINK libspdk_ublk.so 00:01:50.784 SYMLINK libspdk_scsi.so 00:01:51.041 CC lib/vhost/vhost_rpc.o 00:01:51.042 CC lib/vhost/vhost.o 00:01:51.042 CC lib/vhost/vhost_blk.o 00:01:51.042 CC lib/vhost/vhost_scsi.o 00:01:51.042 CC lib/vhost/rte_vhost_user.o 00:01:51.042 CC lib/iscsi/iscsi.o 00:01:51.042 CC lib/iscsi/conn.o 00:01:51.042 CC lib/iscsi/init_grp.o 00:01:51.042 CC lib/iscsi/md5.o 00:01:51.042 CC lib/iscsi/param.o 00:01:51.042 CC lib/iscsi/portal_grp.o 00:01:51.042 CC lib/iscsi/tgt_node.o 00:01:51.042 CC lib/iscsi/iscsi_subsystem.o 00:01:51.042 CC lib/iscsi/iscsi_rpc.o 00:01:51.042 LIB libspdk_ftl.a 00:01:51.042 CC lib/iscsi/task.o 00:01:51.300 SO libspdk_ftl.so.9.0 00:01:51.559 SYMLINK libspdk_ftl.so 00:01:51.817 LIB libspdk_nvmf.a 00:01:51.817 SO libspdk_nvmf.so.19.0 00:01:51.817 LIB libspdk_vhost.a 00:01:52.076 SO libspdk_vhost.so.8.0 00:01:52.076 SYMLINK libspdk_vhost.so 00:01:52.076 SYMLINK libspdk_nvmf.so 00:01:52.076 LIB libspdk_iscsi.a 00:01:52.076 SO libspdk_iscsi.so.8.0 00:01:52.334 SYMLINK libspdk_iscsi.so 00:01:52.901 CC module/vfu_device/vfu_virtio.o 00:01:52.901 CC module/vfu_device/vfu_virtio_scsi.o 00:01:52.901 CC module/vfu_device/vfu_virtio_blk.o 00:01:52.901 CC module/vfu_device/vfu_virtio_rpc.o 00:01:52.901 CC module/env_dpdk/env_dpdk_rpc.o 00:01:52.901 CC module/keyring/linux/keyring.o 00:01:52.901 CC module/keyring/linux/keyring_rpc.o 00:01:52.901 CC module/keyring/file/keyring.o 00:01:52.901 CC module/keyring/file/keyring_rpc.o 00:01:52.901 CC module/accel/ioat/accel_ioat.o 00:01:52.901 CC module/accel/ioat/accel_ioat_rpc.o 00:01:52.901 LIB libspdk_env_dpdk_rpc.a 00:01:52.901 CC module/scheduler/gscheduler/gscheduler.o 00:01:52.901 CC module/blob/bdev/blob_bdev.o 00:01:52.901 CC module/accel/iaa/accel_iaa.o 00:01:52.901 CC module/scheduler/dynamic/scheduler_dynamic.o 00:01:52.901 CC module/accel/iaa/accel_iaa_rpc.o 00:01:52.901 CC module/accel/error/accel_error_rpc.o 00:01:52.901 CC module/accel/error/accel_error.o 00:01:52.901 CC module/scheduler/dpdk_governor/dpdk_governor.o 00:01:52.901 CC module/accel/dsa/accel_dsa.o 00:01:52.901 CC module/accel/dsa/accel_dsa_rpc.o 00:01:52.901 CC module/sock/posix/posix.o 00:01:52.901 SO libspdk_env_dpdk_rpc.so.6.0 00:01:52.901 SYMLINK libspdk_env_dpdk_rpc.so 00:01:53.159 LIB libspdk_keyring_file.a 00:01:53.159 LIB libspdk_keyring_linux.a 00:01:53.159 LIB libspdk_scheduler_gscheduler.a 00:01:53.159 SO libspdk_keyring_file.so.1.0 00:01:53.159 SO libspdk_keyring_linux.so.1.0 00:01:53.159 LIB libspdk_scheduler_dpdk_governor.a 00:01:53.159 LIB libspdk_accel_ioat.a 00:01:53.159 LIB libspdk_accel_error.a 00:01:53.159 SO libspdk_scheduler_gscheduler.so.4.0 00:01:53.159 LIB libspdk_scheduler_dynamic.a 00:01:53.159 LIB libspdk_accel_iaa.a 00:01:53.159 SO libspdk_scheduler_dpdk_governor.so.4.0 00:01:53.159 SO libspdk_accel_ioat.so.6.0 00:01:53.159 SO libspdk_accel_error.so.2.0 00:01:53.159 SYMLINK libspdk_keyring_file.so 00:01:53.159 LIB libspdk_accel_dsa.a 00:01:53.159 SYMLINK libspdk_keyring_linux.so 00:01:53.159 SO libspdk_scheduler_dynamic.so.4.0 00:01:53.159 SO libspdk_accel_iaa.so.3.0 00:01:53.159 LIB libspdk_blob_bdev.a 00:01:53.159 SYMLINK libspdk_scheduler_gscheduler.so 00:01:53.159 SO libspdk_accel_dsa.so.5.0 00:01:53.159 SYMLINK libspdk_scheduler_dpdk_governor.so 00:01:53.159 SO libspdk_blob_bdev.so.11.0 00:01:53.159 SYMLINK libspdk_scheduler_dynamic.so 00:01:53.159 SYMLINK libspdk_accel_error.so 00:01:53.159 SYMLINK libspdk_accel_ioat.so 00:01:53.159 SYMLINK libspdk_accel_iaa.so 00:01:53.159 SYMLINK libspdk_accel_dsa.so 00:01:53.159 LIB libspdk_vfu_device.a 00:01:53.160 SYMLINK libspdk_blob_bdev.so 00:01:53.160 SO libspdk_vfu_device.so.3.0 00:01:53.417 SYMLINK libspdk_vfu_device.so 00:01:53.417 LIB libspdk_sock_posix.a 00:01:53.676 SO libspdk_sock_posix.so.6.0 00:01:53.676 SYMLINK libspdk_sock_posix.so 00:01:53.676 CC module/bdev/aio/bdev_aio_rpc.o 00:01:53.676 CC module/bdev/aio/bdev_aio.o 00:01:53.676 CC module/bdev/error/vbdev_error.o 00:01:53.676 CC module/bdev/error/vbdev_error_rpc.o 00:01:53.676 CC module/bdev/delay/vbdev_delay.o 00:01:53.676 CC module/bdev/delay/vbdev_delay_rpc.o 00:01:53.676 CC module/bdev/passthru/vbdev_passthru.o 00:01:53.676 CC module/bdev/passthru/vbdev_passthru_rpc.o 00:01:53.676 CC module/bdev/zone_block/vbdev_zone_block.o 00:01:53.676 CC module/bdev/zone_block/vbdev_zone_block_rpc.o 00:01:53.676 CC module/bdev/gpt/vbdev_gpt.o 00:01:53.676 CC module/bdev/gpt/gpt.o 00:01:53.676 CC module/bdev/null/bdev_null.o 00:01:53.676 CC module/bdev/iscsi/bdev_iscsi.o 00:01:53.676 CC module/bdev/iscsi/bdev_iscsi_rpc.o 00:01:53.676 CC module/bdev/null/bdev_null_rpc.o 00:01:53.676 CC module/bdev/lvol/vbdev_lvol.o 00:01:53.676 CC module/bdev/lvol/vbdev_lvol_rpc.o 00:01:53.676 CC module/bdev/nvme/bdev_nvme_rpc.o 00:01:53.676 CC module/bdev/nvme/bdev_nvme.o 00:01:53.676 CC module/bdev/nvme/vbdev_opal.o 00:01:53.676 CC module/bdev/nvme/nvme_rpc.o 00:01:53.676 CC module/bdev/nvme/bdev_mdns_client.o 00:01:53.676 CC module/bdev/malloc/bdev_malloc.o 00:01:53.676 CC module/bdev/raid/bdev_raid.o 00:01:53.676 CC module/bdev/nvme/vbdev_opal_rpc.o 00:01:53.676 CC module/bdev/malloc/bdev_malloc_rpc.o 00:01:53.676 CC module/bdev/split/vbdev_split.o 00:01:53.676 CC module/bdev/ftl/bdev_ftl.o 00:01:53.676 CC module/bdev/raid/bdev_raid_rpc.o 00:01:53.676 CC module/bdev/nvme/bdev_nvme_cuse_rpc.o 00:01:53.676 CC module/bdev/split/vbdev_split_rpc.o 00:01:53.676 CC module/bdev/raid/bdev_raid_sb.o 00:01:53.676 CC module/bdev/raid/raid0.o 00:01:53.676 CC module/bdev/raid/raid1.o 00:01:53.676 CC module/bdev/ftl/bdev_ftl_rpc.o 00:01:53.676 CC module/bdev/raid/concat.o 00:01:53.676 CC module/blobfs/bdev/blobfs_bdev.o 00:01:53.676 CC module/blobfs/bdev/blobfs_bdev_rpc.o 00:01:53.676 CC module/bdev/virtio/bdev_virtio_scsi.o 00:01:53.676 CC module/bdev/virtio/bdev_virtio_blk.o 00:01:53.676 CC module/bdev/virtio/bdev_virtio_rpc.o 00:01:53.935 LIB libspdk_blobfs_bdev.a 00:01:53.935 LIB libspdk_bdev_error.a 00:01:53.935 LIB libspdk_bdev_split.a 00:01:53.935 SO libspdk_blobfs_bdev.so.6.0 00:01:53.935 LIB libspdk_bdev_null.a 00:01:53.935 SO libspdk_bdev_error.so.6.0 00:01:53.935 SO libspdk_bdev_split.so.6.0 00:01:53.935 LIB libspdk_bdev_passthru.a 00:01:53.935 SO libspdk_bdev_null.so.6.0 00:01:53.935 LIB libspdk_bdev_gpt.a 00:01:53.935 LIB libspdk_bdev_aio.a 00:01:53.935 LIB libspdk_bdev_ftl.a 00:01:53.935 SYMLINK libspdk_blobfs_bdev.so 00:01:53.935 SO libspdk_bdev_passthru.so.6.0 00:01:53.935 SO libspdk_bdev_ftl.so.6.0 00:01:53.935 SO libspdk_bdev_gpt.so.6.0 00:01:53.935 SO libspdk_bdev_aio.so.6.0 00:01:53.935 SYMLINK libspdk_bdev_error.so 00:01:53.935 SYMLINK libspdk_bdev_split.so 00:01:53.935 LIB libspdk_bdev_zone_block.a 00:01:53.935 SYMLINK libspdk_bdev_null.so 00:01:53.935 LIB libspdk_bdev_iscsi.a 00:01:53.935 SO libspdk_bdev_zone_block.so.6.0 00:01:54.193 LIB libspdk_bdev_delay.a 00:01:54.193 SYMLINK libspdk_bdev_ftl.so 00:01:54.193 SYMLINK libspdk_bdev_passthru.so 00:01:54.193 SYMLINK libspdk_bdev_gpt.so 00:01:54.193 LIB libspdk_bdev_malloc.a 00:01:54.193 SYMLINK libspdk_bdev_aio.so 00:01:54.193 SO libspdk_bdev_delay.so.6.0 00:01:54.193 SO libspdk_bdev_iscsi.so.6.0 00:01:54.193 SO libspdk_bdev_malloc.so.6.0 00:01:54.193 SYMLINK libspdk_bdev_zone_block.so 00:01:54.193 SYMLINK libspdk_bdev_delay.so 00:01:54.193 SYMLINK libspdk_bdev_iscsi.so 00:01:54.193 SYMLINK libspdk_bdev_malloc.so 00:01:54.193 LIB libspdk_bdev_lvol.a 00:01:54.193 LIB libspdk_bdev_virtio.a 00:01:54.193 SO libspdk_bdev_lvol.so.6.0 00:01:54.193 SO libspdk_bdev_virtio.so.6.0 00:01:54.193 SYMLINK libspdk_bdev_virtio.so 00:01:54.193 SYMLINK libspdk_bdev_lvol.so 00:01:54.452 LIB libspdk_bdev_raid.a 00:01:54.452 SO libspdk_bdev_raid.so.6.0 00:01:54.710 SYMLINK libspdk_bdev_raid.so 00:01:55.277 LIB libspdk_bdev_nvme.a 00:01:55.277 SO libspdk_bdev_nvme.so.7.0 00:01:55.535 SYMLINK libspdk_bdev_nvme.so 00:01:56.101 CC module/event/subsystems/vhost_blk/vhost_blk.o 00:01:56.101 CC module/event/subsystems/sock/sock.o 00:01:56.101 CC module/event/subsystems/vmd/vmd.o 00:01:56.101 CC module/event/subsystems/vmd/vmd_rpc.o 00:01:56.101 CC module/event/subsystems/scheduler/scheduler.o 00:01:56.101 CC module/event/subsystems/iobuf/iobuf.o 00:01:56.101 CC module/event/subsystems/iobuf/iobuf_rpc.o 00:01:56.101 CC module/event/subsystems/keyring/keyring.o 00:01:56.101 CC module/event/subsystems/vfu_tgt/vfu_tgt.o 00:01:56.101 LIB libspdk_event_vhost_blk.a 00:01:56.360 LIB libspdk_event_vmd.a 00:01:56.360 LIB libspdk_event_sock.a 00:01:56.360 LIB libspdk_event_scheduler.a 00:01:56.360 SO libspdk_event_vhost_blk.so.3.0 00:01:56.360 LIB libspdk_event_vfu_tgt.a 00:01:56.360 LIB libspdk_event_keyring.a 00:01:56.360 LIB libspdk_event_iobuf.a 00:01:56.360 SO libspdk_event_keyring.so.1.0 00:01:56.360 SO libspdk_event_vfu_tgt.so.3.0 00:01:56.360 SO libspdk_event_scheduler.so.4.0 00:01:56.360 SO libspdk_event_sock.so.5.0 00:01:56.360 SO libspdk_event_vmd.so.6.0 00:01:56.360 SYMLINK libspdk_event_vhost_blk.so 00:01:56.360 SO libspdk_event_iobuf.so.3.0 00:01:56.360 SYMLINK libspdk_event_sock.so 00:01:56.360 SYMLINK libspdk_event_keyring.so 00:01:56.360 SYMLINK libspdk_event_scheduler.so 00:01:56.360 SYMLINK libspdk_event_vfu_tgt.so 00:01:56.360 SYMLINK libspdk_event_vmd.so 00:01:56.360 SYMLINK libspdk_event_iobuf.so 00:01:56.654 CC module/event/subsystems/accel/accel.o 00:01:56.654 LIB libspdk_event_accel.a 00:01:56.913 SO libspdk_event_accel.so.6.0 00:01:56.913 SYMLINK libspdk_event_accel.so 00:01:57.172 CC module/event/subsystems/bdev/bdev.o 00:01:57.432 LIB libspdk_event_bdev.a 00:01:57.432 SO libspdk_event_bdev.so.6.0 00:01:57.432 SYMLINK libspdk_event_bdev.so 00:01:57.690 CC module/event/subsystems/nvmf/nvmf_rpc.o 00:01:57.690 CC module/event/subsystems/nvmf/nvmf_tgt.o 00:01:57.690 CC module/event/subsystems/nbd/nbd.o 00:01:57.690 CC module/event/subsystems/scsi/scsi.o 00:01:57.690 CC module/event/subsystems/ublk/ublk.o 00:01:57.948 LIB libspdk_event_nbd.a 00:01:57.948 LIB libspdk_event_scsi.a 00:01:57.948 LIB libspdk_event_ublk.a 00:01:57.948 SO libspdk_event_nbd.so.6.0 00:01:57.948 LIB libspdk_event_nvmf.a 00:01:57.948 SO libspdk_event_scsi.so.6.0 00:01:57.948 SO libspdk_event_ublk.so.3.0 00:01:57.948 SO libspdk_event_nvmf.so.6.0 00:01:57.948 SYMLINK libspdk_event_nbd.so 00:01:57.948 SYMLINK libspdk_event_scsi.so 00:01:57.948 SYMLINK libspdk_event_ublk.so 00:01:57.948 SYMLINK libspdk_event_nvmf.so 00:01:58.206 CC module/event/subsystems/vhost_scsi/vhost_scsi.o 00:01:58.206 CC module/event/subsystems/iscsi/iscsi.o 00:01:58.465 LIB libspdk_event_vhost_scsi.a 00:01:58.465 LIB libspdk_event_iscsi.a 00:01:58.465 SO libspdk_event_vhost_scsi.so.3.0 00:01:58.465 SO libspdk_event_iscsi.so.6.0 00:01:58.465 SYMLINK libspdk_event_vhost_scsi.so 00:01:58.465 SYMLINK libspdk_event_iscsi.so 00:01:58.723 SO libspdk.so.6.0 00:01:58.723 SYMLINK libspdk.so 00:01:58.982 CC app/trace_record/trace_record.o 00:01:58.982 CC app/spdk_top/spdk_top.o 00:01:58.982 CC app/spdk_lspci/spdk_lspci.o 00:01:58.982 CC app/spdk_nvme_identify/identify.o 00:01:58.982 CXX app/trace/trace.o 00:01:58.982 CC app/spdk_nvme_perf/perf.o 00:01:58.982 CC test/rpc_client/rpc_client_test.o 00:01:58.982 CC app/spdk_nvme_discover/discovery_aer.o 00:01:58.982 CC examples/interrupt_tgt/interrupt_tgt.o 00:01:58.982 TEST_HEADER include/spdk/accel_module.h 00:01:58.982 TEST_HEADER include/spdk/accel.h 00:01:58.982 TEST_HEADER include/spdk/assert.h 00:01:58.982 TEST_HEADER include/spdk/barrier.h 00:01:58.982 TEST_HEADER include/spdk/bdev.h 00:01:58.982 TEST_HEADER include/spdk/base64.h 00:01:58.982 TEST_HEADER include/spdk/bdev_zone.h 00:01:58.982 TEST_HEADER include/spdk/bit_array.h 00:01:58.982 TEST_HEADER include/spdk/bdev_module.h 00:01:58.982 TEST_HEADER include/spdk/bit_pool.h 00:01:58.982 TEST_HEADER include/spdk/blob_bdev.h 00:01:58.982 TEST_HEADER include/spdk/blobfs_bdev.h 00:01:58.982 TEST_HEADER include/spdk/blob.h 00:01:58.982 TEST_HEADER include/spdk/conf.h 00:01:58.982 TEST_HEADER include/spdk/blobfs.h 00:01:58.982 TEST_HEADER include/spdk/config.h 00:01:58.982 TEST_HEADER include/spdk/cpuset.h 00:01:58.982 TEST_HEADER include/spdk/crc32.h 00:01:58.982 TEST_HEADER include/spdk/crc16.h 00:01:58.982 TEST_HEADER include/spdk/crc64.h 00:01:58.982 TEST_HEADER include/spdk/endian.h 00:01:58.982 TEST_HEADER include/spdk/dif.h 00:01:58.982 TEST_HEADER include/spdk/dma.h 00:01:58.982 TEST_HEADER include/spdk/env.h 00:01:58.982 TEST_HEADER include/spdk/env_dpdk.h 00:01:58.982 CC app/spdk_dd/spdk_dd.o 00:01:58.982 TEST_HEADER include/spdk/fd.h 00:01:58.982 TEST_HEADER include/spdk/event.h 00:01:58.982 TEST_HEADER include/spdk/file.h 00:01:58.982 TEST_HEADER include/spdk/fd_group.h 00:01:58.982 CC app/iscsi_tgt/iscsi_tgt.o 00:01:58.982 TEST_HEADER include/spdk/gpt_spec.h 00:01:58.982 TEST_HEADER include/spdk/hexlify.h 00:01:58.982 TEST_HEADER include/spdk/histogram_data.h 00:01:58.982 TEST_HEADER include/spdk/ftl.h 00:01:58.982 TEST_HEADER include/spdk/idxd.h 00:01:58.982 TEST_HEADER include/spdk/init.h 00:01:58.982 TEST_HEADER include/spdk/idxd_spec.h 00:01:58.982 TEST_HEADER include/spdk/ioat_spec.h 00:01:58.982 TEST_HEADER include/spdk/iscsi_spec.h 00:01:58.982 TEST_HEADER include/spdk/ioat.h 00:01:58.982 TEST_HEADER include/spdk/keyring.h 00:01:58.982 TEST_HEADER include/spdk/json.h 00:01:58.982 CC app/nvmf_tgt/nvmf_main.o 00:01:58.982 TEST_HEADER include/spdk/keyring_module.h 00:01:58.982 TEST_HEADER include/spdk/log.h 00:01:58.982 TEST_HEADER include/spdk/jsonrpc.h 00:01:58.982 TEST_HEADER include/spdk/likely.h 00:01:58.982 TEST_HEADER include/spdk/lvol.h 00:01:58.982 TEST_HEADER include/spdk/memory.h 00:01:58.982 TEST_HEADER include/spdk/mmio.h 00:01:58.982 TEST_HEADER include/spdk/nbd.h 00:01:58.982 TEST_HEADER include/spdk/notify.h 00:01:58.982 TEST_HEADER include/spdk/nvme.h 00:01:58.982 TEST_HEADER include/spdk/nvme_ocssd.h 00:01:58.982 TEST_HEADER include/spdk/nvme_ocssd_spec.h 00:01:58.982 TEST_HEADER include/spdk/nvme_intel.h 00:01:58.982 TEST_HEADER include/spdk/nvme_zns.h 00:01:58.982 TEST_HEADER include/spdk/nvme_spec.h 00:01:58.982 TEST_HEADER include/spdk/nvmf_cmd.h 00:01:58.982 TEST_HEADER include/spdk/nvmf_fc_spec.h 00:01:58.982 TEST_HEADER include/spdk/nvmf_spec.h 00:01:58.982 TEST_HEADER include/spdk/nvmf.h 00:01:58.982 TEST_HEADER include/spdk/opal.h 00:01:58.982 TEST_HEADER include/spdk/opal_spec.h 00:01:58.982 TEST_HEADER include/spdk/pci_ids.h 00:01:58.982 TEST_HEADER include/spdk/nvmf_transport.h 00:01:58.982 TEST_HEADER include/spdk/queue.h 00:01:58.982 TEST_HEADER include/spdk/rpc.h 00:01:58.982 TEST_HEADER include/spdk/pipe.h 00:01:58.982 TEST_HEADER include/spdk/reduce.h 00:01:58.982 CC app/spdk_tgt/spdk_tgt.o 00:01:58.982 TEST_HEADER include/spdk/scheduler.h 00:01:58.982 TEST_HEADER include/spdk/scsi_spec.h 00:01:58.982 TEST_HEADER include/spdk/sock.h 00:01:58.982 TEST_HEADER include/spdk/scsi.h 00:01:58.982 TEST_HEADER include/spdk/stdinc.h 00:01:58.982 TEST_HEADER include/spdk/trace.h 00:01:58.982 TEST_HEADER include/spdk/string.h 00:01:58.982 TEST_HEADER include/spdk/thread.h 00:01:58.982 TEST_HEADER include/spdk/trace_parser.h 00:01:58.982 TEST_HEADER include/spdk/ublk.h 00:01:58.982 TEST_HEADER include/spdk/tree.h 00:01:58.982 TEST_HEADER include/spdk/util.h 00:01:58.982 TEST_HEADER include/spdk/uuid.h 00:01:58.982 TEST_HEADER include/spdk/vfio_user_spec.h 00:01:58.982 TEST_HEADER include/spdk/version.h 00:01:58.982 TEST_HEADER include/spdk/vfio_user_pci.h 00:01:58.982 TEST_HEADER include/spdk/vhost.h 00:01:58.982 TEST_HEADER include/spdk/vmd.h 00:01:58.982 TEST_HEADER include/spdk/xor.h 00:01:58.982 TEST_HEADER include/spdk/zipf.h 00:01:58.982 CXX test/cpp_headers/accel_module.o 00:01:58.982 CXX test/cpp_headers/accel.o 00:01:58.982 CXX test/cpp_headers/assert.o 00:01:58.982 CXX test/cpp_headers/barrier.o 00:01:58.982 CXX test/cpp_headers/bdev.o 00:01:58.982 CXX test/cpp_headers/bdev_module.o 00:01:58.982 CXX test/cpp_headers/base64.o 00:01:58.982 CXX test/cpp_headers/bdev_zone.o 00:01:58.982 CXX test/cpp_headers/bit_array.o 00:01:58.982 CXX test/cpp_headers/bit_pool.o 00:01:58.982 CXX test/cpp_headers/blob_bdev.o 00:01:58.982 CXX test/cpp_headers/blobfs_bdev.o 00:01:58.982 CXX test/cpp_headers/blobfs.o 00:01:58.982 CXX test/cpp_headers/conf.o 00:01:58.982 CXX test/cpp_headers/cpuset.o 00:01:58.982 CXX test/cpp_headers/blob.o 00:01:58.982 CXX test/cpp_headers/config.o 00:01:58.982 CXX test/cpp_headers/crc16.o 00:01:58.982 CXX test/cpp_headers/crc32.o 00:01:58.982 CXX test/cpp_headers/crc64.o 00:01:58.982 CXX test/cpp_headers/dif.o 00:01:58.982 CXX test/cpp_headers/dma.o 00:01:58.982 CXX test/cpp_headers/endian.o 00:01:58.982 CXX test/cpp_headers/env.o 00:01:58.982 CC examples/ioat/perf/perf.o 00:01:58.982 CXX test/cpp_headers/env_dpdk.o 00:01:58.982 CXX test/cpp_headers/event.o 00:01:58.982 CXX test/cpp_headers/fd_group.o 00:01:58.982 CXX test/cpp_headers/fd.o 00:01:58.982 CXX test/cpp_headers/file.o 00:01:59.252 CXX test/cpp_headers/ftl.o 00:01:59.252 CC examples/ioat/verify/verify.o 00:01:59.252 CXX test/cpp_headers/histogram_data.o 00:01:59.252 CXX test/cpp_headers/gpt_spec.o 00:01:59.252 CXX test/cpp_headers/hexlify.o 00:01:59.252 CXX test/cpp_headers/idxd_spec.o 00:01:59.252 CXX test/cpp_headers/idxd.o 00:01:59.252 CXX test/cpp_headers/ioat.o 00:01:59.252 CXX test/cpp_headers/init.o 00:01:59.252 CC examples/util/zipf/zipf.o 00:01:59.252 CXX test/cpp_headers/iscsi_spec.o 00:01:59.252 CXX test/cpp_headers/ioat_spec.o 00:01:59.252 CXX test/cpp_headers/json.o 00:01:59.252 CXX test/cpp_headers/jsonrpc.o 00:01:59.252 CXX test/cpp_headers/keyring_module.o 00:01:59.252 CXX test/cpp_headers/keyring.o 00:01:59.252 CXX test/cpp_headers/log.o 00:01:59.252 CXX test/cpp_headers/likely.o 00:01:59.252 CXX test/cpp_headers/lvol.o 00:01:59.252 CXX test/cpp_headers/memory.o 00:01:59.252 CXX test/cpp_headers/mmio.o 00:01:59.252 CXX test/cpp_headers/nbd.o 00:01:59.252 CXX test/cpp_headers/notify.o 00:01:59.252 CXX test/cpp_headers/nvme.o 00:01:59.253 CXX test/cpp_headers/nvme_intel.o 00:01:59.253 CXX test/cpp_headers/nvme_ocssd.o 00:01:59.253 CXX test/cpp_headers/nvme_ocssd_spec.o 00:01:59.253 CXX test/cpp_headers/nvme_spec.o 00:01:59.253 CXX test/cpp_headers/nvme_zns.o 00:01:59.253 CXX test/cpp_headers/nvmf.o 00:01:59.253 CXX test/cpp_headers/nvmf_cmd.o 00:01:59.253 CXX test/cpp_headers/nvmf_fc_spec.o 00:01:59.253 CC app/fio/nvme/fio_plugin.o 00:01:59.253 CXX test/cpp_headers/nvmf_spec.o 00:01:59.253 CC test/app/jsoncat/jsoncat.o 00:01:59.253 CXX test/cpp_headers/nvmf_transport.o 00:01:59.253 CXX test/cpp_headers/opal.o 00:01:59.253 CXX test/cpp_headers/opal_spec.o 00:01:59.253 CXX test/cpp_headers/pci_ids.o 00:01:59.253 CXX test/cpp_headers/pipe.o 00:01:59.253 CC test/env/pci/pci_ut.o 00:01:59.253 CC test/app/stub/stub.o 00:01:59.253 CC test/env/env_dpdk_post_init/env_dpdk_post_init.o 00:01:59.253 CC test/app/histogram_perf/histogram_perf.o 00:01:59.253 CC test/env/memory/memory_ut.o 00:01:59.253 CC test/thread/poller_perf/poller_perf.o 00:01:59.253 CC test/env/vtophys/vtophys.o 00:01:59.253 CC app/fio/bdev/fio_plugin.o 00:01:59.253 CXX test/cpp_headers/queue.o 00:01:59.253 CC test/app/bdev_svc/bdev_svc.o 00:01:59.253 CC test/dma/test_dma/test_dma.o 00:01:59.253 LINK spdk_lspci 00:01:59.253 CXX test/cpp_headers/reduce.o 00:01:59.253 LINK interrupt_tgt 00:01:59.522 LINK spdk_nvme_discover 00:01:59.522 CC test/app/fuzz/iscsi_fuzz/iscsi_fuzz.o 00:01:59.522 CC test/env/mem_callbacks/mem_callbacks.o 00:01:59.522 CC test/app/fuzz/nvme_fuzz/nvme_fuzz.o 00:01:59.522 CC test/app/fuzz/vhost_fuzz/vhost_fuzz_rpc.o 00:01:59.522 LINK rpc_client_test 00:01:59.522 LINK jsoncat 00:01:59.781 LINK histogram_perf 00:01:59.781 LINK ioat_perf 00:01:59.781 LINK nvmf_tgt 00:01:59.781 LINK iscsi_tgt 00:01:59.781 CC test/app/fuzz/vhost_fuzz/vhost_fuzz.o 00:01:59.781 CXX test/cpp_headers/rpc.o 00:01:59.781 CXX test/cpp_headers/scheduler.o 00:01:59.781 CXX test/cpp_headers/scsi.o 00:01:59.781 CXX test/cpp_headers/scsi_spec.o 00:01:59.781 CXX test/cpp_headers/sock.o 00:01:59.781 CXX test/cpp_headers/stdinc.o 00:01:59.781 CXX test/cpp_headers/string.o 00:01:59.781 CXX test/cpp_headers/thread.o 00:01:59.781 CXX test/cpp_headers/trace.o 00:01:59.781 LINK zipf 00:01:59.781 CXX test/cpp_headers/trace_parser.o 00:01:59.781 CXX test/cpp_headers/tree.o 00:01:59.781 CXX test/cpp_headers/ublk.o 00:01:59.781 LINK bdev_svc 00:01:59.781 CXX test/cpp_headers/util.o 00:01:59.781 CXX test/cpp_headers/uuid.o 00:01:59.781 CXX test/cpp_headers/version.o 00:01:59.781 CXX test/cpp_headers/vfio_user_pci.o 00:01:59.781 CXX test/cpp_headers/vfio_user_spec.o 00:01:59.781 LINK spdk_trace_record 00:01:59.781 CXX test/cpp_headers/vhost.o 00:01:59.781 CXX test/cpp_headers/vmd.o 00:01:59.781 CXX test/cpp_headers/xor.o 00:01:59.781 CXX test/cpp_headers/zipf.o 00:01:59.781 LINK spdk_tgt 00:01:59.781 LINK spdk_dd 00:01:59.781 LINK vtophys 00:01:59.781 LINK env_dpdk_post_init 00:01:59.781 LINK poller_perf 00:01:59.781 LINK stub 00:01:59.781 LINK verify 00:01:59.781 LINK spdk_trace 00:02:00.039 LINK test_dma 00:02:00.039 LINK pci_ut 00:02:00.039 LINK spdk_nvme 00:02:00.039 LINK spdk_bdev 00:02:00.039 LINK nvme_fuzz 00:02:00.296 LINK spdk_top 00:02:00.296 LINK mem_callbacks 00:02:00.296 CC examples/idxd/perf/perf.o 00:02:00.296 CC app/vhost/vhost.o 00:02:00.296 CC examples/vmd/lsvmd/lsvmd.o 00:02:00.296 CC examples/vmd/led/led.o 00:02:00.296 LINK vhost_fuzz 00:02:00.296 CC examples/sock/hello_world/hello_sock.o 00:02:00.296 CC test/event/reactor_perf/reactor_perf.o 00:02:00.296 CC test/event/reactor/reactor.o 00:02:00.296 CC examples/thread/thread/thread_ex.o 00:02:00.296 CC test/event/event_perf/event_perf.o 00:02:00.296 CC test/event/app_repeat/app_repeat.o 00:02:00.296 LINK spdk_nvme_perf 00:02:00.296 CC test/event/scheduler/scheduler.o 00:02:00.296 LINK spdk_nvme_identify 00:02:00.296 LINK led 00:02:00.296 CC test/nvme/sgl/sgl.o 00:02:00.296 CC test/nvme/overhead/overhead.o 00:02:00.296 CC test/nvme/e2edp/nvme_dp.o 00:02:00.296 CC test/nvme/boot_partition/boot_partition.o 00:02:00.296 CC test/nvme/doorbell_aers/doorbell_aers.o 00:02:00.296 CC test/nvme/reset/reset.o 00:02:00.297 CC test/nvme/compliance/nvme_compliance.o 00:02:00.297 CC test/nvme/reserve/reserve.o 00:02:00.297 LINK lsvmd 00:02:00.297 CC test/nvme/err_injection/err_injection.o 00:02:00.297 CC test/nvme/fdp/fdp.o 00:02:00.297 CC test/nvme/fused_ordering/fused_ordering.o 00:02:00.297 CC test/nvme/simple_copy/simple_copy.o 00:02:00.297 CC test/nvme/connect_stress/connect_stress.o 00:02:00.297 CC test/nvme/cuse/cuse.o 00:02:00.555 CC test/accel/dif/dif.o 00:02:00.555 CC test/nvme/aer/aer.o 00:02:00.555 CC test/nvme/startup/startup.o 00:02:00.555 LINK reactor 00:02:00.555 CC test/blobfs/mkfs/mkfs.o 00:02:00.555 LINK reactor_perf 00:02:00.555 LINK vhost 00:02:00.555 LINK event_perf 00:02:00.555 LINK app_repeat 00:02:00.555 LINK hello_sock 00:02:00.555 CC test/lvol/esnap/esnap.o 00:02:00.555 LINK idxd_perf 00:02:00.555 LINK thread 00:02:00.555 LINK scheduler 00:02:00.555 LINK boot_partition 00:02:00.555 LINK connect_stress 00:02:00.555 LINK doorbell_aers 00:02:00.555 LINK startup 00:02:00.555 LINK err_injection 00:02:00.555 LINK fused_ordering 00:02:00.555 LINK reserve 00:02:00.555 LINK simple_copy 00:02:00.555 LINK memory_ut 00:02:00.555 LINK mkfs 00:02:00.555 LINK sgl 00:02:00.555 LINK overhead 00:02:00.555 LINK reset 00:02:00.555 LINK nvme_dp 00:02:00.813 LINK aer 00:02:00.813 LINK nvme_compliance 00:02:00.813 LINK fdp 00:02:00.813 LINK dif 00:02:00.813 CC examples/nvme/arbitration/arbitration.o 00:02:00.813 CC examples/nvme/hotplug/hotplug.o 00:02:00.813 CC examples/nvme/hello_world/hello_world.o 00:02:01.072 CC examples/nvme/reconnect/reconnect.o 00:02:01.072 CC examples/nvme/cmb_copy/cmb_copy.o 00:02:01.072 CC examples/nvme/abort/abort.o 00:02:01.072 CC examples/nvme/nvme_manage/nvme_manage.o 00:02:01.072 CC examples/nvme/pmr_persistence/pmr_persistence.o 00:02:01.072 CC examples/accel/perf/accel_perf.o 00:02:01.072 CC examples/blob/cli/blobcli.o 00:02:01.072 CC examples/blob/hello_world/hello_blob.o 00:02:01.072 LINK iscsi_fuzz 00:02:01.072 LINK cmb_copy 00:02:01.072 LINK pmr_persistence 00:02:01.072 LINK hello_world 00:02:01.072 LINK hotplug 00:02:01.330 LINK arbitration 00:02:01.330 LINK hello_blob 00:02:01.330 LINK reconnect 00:02:01.330 LINK abort 00:02:01.330 CC test/bdev/bdevio/bdevio.o 00:02:01.330 LINK nvme_manage 00:02:01.330 LINK accel_perf 00:02:01.330 LINK cuse 00:02:01.330 LINK blobcli 00:02:01.589 LINK bdevio 00:02:01.848 CC examples/bdev/bdevperf/bdevperf.o 00:02:01.848 CC examples/bdev/hello_world/hello_bdev.o 00:02:02.107 LINK hello_bdev 00:02:02.366 LINK bdevperf 00:02:02.935 CC examples/nvmf/nvmf/nvmf.o 00:02:03.194 LINK nvmf 00:02:04.131 LINK esnap 00:02:04.131 00:02:04.131 real 0m43.187s 00:02:04.131 user 6m30.492s 00:02:04.131 sys 3m20.395s 00:02:04.131 17:47:57 make -- common/autotest_common.sh@1124 -- $ xtrace_disable 00:02:04.131 17:47:57 make -- common/autotest_common.sh@10 -- $ set +x 00:02:04.131 ************************************ 00:02:04.131 END TEST make 00:02:04.131 ************************************ 00:02:04.390 17:47:57 -- common/autotest_common.sh@1142 -- $ return 0 00:02:04.390 17:47:57 -- spdk/autobuild.sh@1 -- $ stop_monitor_resources 00:02:04.390 17:47:57 -- pm/common@29 -- $ signal_monitor_resources TERM 00:02:04.390 17:47:57 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:02:04.390 17:47:57 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:04.390 17:47:57 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-cpu-load.pid ]] 00:02:04.390 17:47:57 -- pm/common@44 -- $ pid=298595 00:02:04.390 17:47:57 -- pm/common@50 -- $ kill -TERM 298595 00:02:04.390 17:47:57 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:04.390 17:47:57 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-vmstat.pid ]] 00:02:04.390 17:47:57 -- pm/common@44 -- $ pid=298596 00:02:04.390 17:47:57 -- pm/common@50 -- $ kill -TERM 298596 00:02:04.390 17:47:57 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:04.390 17:47:57 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-cpu-temp.pid ]] 00:02:04.390 17:47:57 -- pm/common@44 -- $ pid=298597 00:02:04.390 17:47:57 -- pm/common@50 -- $ kill -TERM 298597 00:02:04.390 17:47:57 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:02:04.390 17:47:57 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-bmc-pm.pid ]] 00:02:04.390 17:47:57 -- pm/common@44 -- $ pid=298619 00:02:04.390 17:47:57 -- pm/common@50 -- $ sudo -E kill -TERM 298619 00:02:04.390 17:47:57 -- spdk/autotest.sh@25 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:02:04.390 17:47:57 -- nvmf/common.sh@7 -- # uname -s 00:02:04.390 17:47:57 -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:02:04.390 17:47:57 -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:02:04.390 17:47:57 -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:02:04.390 17:47:57 -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:02:04.390 17:47:57 -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:02:04.390 17:47:57 -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:02:04.391 17:47:57 -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:02:04.391 17:47:57 -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:02:04.391 17:47:57 -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:02:04.391 17:47:57 -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:02:04.391 17:47:58 -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:02:04.391 17:47:58 -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:02:04.391 17:47:58 -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:02:04.391 17:47:58 -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:02:04.391 17:47:58 -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:02:04.391 17:47:58 -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:02:04.391 17:47:58 -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:02:04.391 17:47:58 -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:02:04.391 17:47:58 -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:02:04.391 17:47:58 -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:02:04.391 17:47:58 -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:04.391 17:47:58 -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:04.391 17:47:58 -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:04.391 17:47:58 -- paths/export.sh@5 -- # export PATH 00:02:04.391 17:47:58 -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:02:04.391 17:47:58 -- nvmf/common.sh@47 -- # : 0 00:02:04.391 17:47:58 -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:02:04.391 17:47:58 -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:02:04.391 17:47:58 -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:02:04.391 17:47:58 -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:02:04.391 17:47:58 -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:02:04.391 17:47:58 -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:02:04.391 17:47:58 -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:02:04.391 17:47:58 -- nvmf/common.sh@51 -- # have_pci_nics=0 00:02:04.391 17:47:58 -- spdk/autotest.sh@27 -- # '[' 0 -ne 0 ']' 00:02:04.391 17:47:58 -- spdk/autotest.sh@32 -- # uname -s 00:02:04.391 17:47:58 -- spdk/autotest.sh@32 -- # '[' Linux = Linux ']' 00:02:04.391 17:47:58 -- spdk/autotest.sh@33 -- # old_core_pattern='|/usr/lib/systemd/systemd-coredump %P %u %g %s %t %c %h' 00:02:04.391 17:47:58 -- spdk/autotest.sh@34 -- # mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/coredumps 00:02:04.391 17:47:58 -- spdk/autotest.sh@39 -- # echo '|/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/core-collector.sh %P %s %t' 00:02:04.391 17:47:58 -- spdk/autotest.sh@40 -- # echo /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/coredumps 00:02:04.391 17:47:58 -- spdk/autotest.sh@44 -- # modprobe nbd 00:02:04.391 17:47:58 -- spdk/autotest.sh@46 -- # type -P udevadm 00:02:04.391 17:47:58 -- spdk/autotest.sh@46 -- # udevadm=/usr/sbin/udevadm 00:02:04.391 17:47:58 -- spdk/autotest.sh@47 -- # /usr/sbin/udevadm monitor --property 00:02:04.391 17:47:58 -- spdk/autotest.sh@48 -- # udevadm_pid=357387 00:02:04.391 17:47:58 -- spdk/autotest.sh@53 -- # start_monitor_resources 00:02:04.391 17:47:58 -- pm/common@17 -- # local monitor 00:02:04.391 17:47:58 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:02:04.391 17:47:58 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:02:04.391 17:47:58 -- pm/common@21 -- # date +%s 00:02:04.391 17:47:58 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:02:04.391 17:47:58 -- pm/common@19 -- # for monitor in "${MONITOR_RESOURCES[@]}" 00:02:04.391 17:47:58 -- pm/common@21 -- # date +%s 00:02:04.391 17:47:58 -- pm/common@25 -- # sleep 1 00:02:04.391 17:47:58 -- pm/common@21 -- # date +%s 00:02:04.391 17:47:58 -- pm/common@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-load -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1721058478 00:02:04.391 17:47:58 -- pm/common@21 -- # date +%s 00:02:04.391 17:47:58 -- pm/common@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-vmstat -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1721058478 00:02:04.391 17:47:58 -- pm/common@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-temp -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1721058478 00:02:04.391 17:47:58 -- pm/common@21 -- # sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-bmc-pm -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autotest.sh.1721058478 00:02:04.391 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1721058478_collect-vmstat.pm.log 00:02:04.391 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1721058478_collect-cpu-load.pm.log 00:02:04.391 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1721058478_collect-cpu-temp.pm.log 00:02:04.391 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autotest.sh.1721058478_collect-bmc-pm.bmc.pm.log 00:02:05.327 17:47:59 -- spdk/autotest.sh@55 -- # trap 'autotest_cleanup || :; exit 1' SIGINT SIGTERM EXIT 00:02:05.327 17:47:59 -- spdk/autotest.sh@57 -- # timing_enter autotest 00:02:05.327 17:47:59 -- common/autotest_common.sh@722 -- # xtrace_disable 00:02:05.327 17:47:59 -- common/autotest_common.sh@10 -- # set +x 00:02:05.327 17:47:59 -- spdk/autotest.sh@59 -- # create_test_list 00:02:05.327 17:47:59 -- common/autotest_common.sh@746 -- # xtrace_disable 00:02:05.327 17:47:59 -- common/autotest_common.sh@10 -- # set +x 00:02:05.586 17:47:59 -- spdk/autotest.sh@61 -- # dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/autotest.sh 00:02:05.586 17:47:59 -- spdk/autotest.sh@61 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:02:05.586 17:47:59 -- spdk/autotest.sh@61 -- # src=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:02:05.586 17:47:59 -- spdk/autotest.sh@62 -- # out=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output 00:02:05.586 17:47:59 -- spdk/autotest.sh@63 -- # cd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:02:05.586 17:47:59 -- spdk/autotest.sh@65 -- # freebsd_update_contigmem_mod 00:02:05.586 17:47:59 -- common/autotest_common.sh@1455 -- # uname 00:02:05.586 17:47:59 -- common/autotest_common.sh@1455 -- # '[' Linux = FreeBSD ']' 00:02:05.586 17:47:59 -- spdk/autotest.sh@66 -- # freebsd_set_maxsock_buf 00:02:05.586 17:47:59 -- common/autotest_common.sh@1475 -- # uname 00:02:05.586 17:47:59 -- common/autotest_common.sh@1475 -- # [[ Linux = FreeBSD ]] 00:02:05.586 17:47:59 -- spdk/autotest.sh@71 -- # grep CC_TYPE mk/cc.mk 00:02:05.586 17:47:59 -- spdk/autotest.sh@71 -- # CC_TYPE=CC_TYPE=gcc 00:02:05.586 17:47:59 -- spdk/autotest.sh@72 -- # hash lcov 00:02:05.586 17:47:59 -- spdk/autotest.sh@72 -- # [[ CC_TYPE=gcc == *\c\l\a\n\g* ]] 00:02:05.586 17:47:59 -- spdk/autotest.sh@80 -- # export 'LCOV_OPTS= 00:02:05.586 --rc lcov_branch_coverage=1 00:02:05.586 --rc lcov_function_coverage=1 00:02:05.586 --rc genhtml_branch_coverage=1 00:02:05.586 --rc genhtml_function_coverage=1 00:02:05.586 --rc genhtml_legend=1 00:02:05.586 --rc geninfo_all_blocks=1 00:02:05.586 ' 00:02:05.586 17:47:59 -- spdk/autotest.sh@80 -- # LCOV_OPTS=' 00:02:05.586 --rc lcov_branch_coverage=1 00:02:05.586 --rc lcov_function_coverage=1 00:02:05.586 --rc genhtml_branch_coverage=1 00:02:05.586 --rc genhtml_function_coverage=1 00:02:05.586 --rc genhtml_legend=1 00:02:05.586 --rc geninfo_all_blocks=1 00:02:05.586 ' 00:02:05.586 17:47:59 -- spdk/autotest.sh@81 -- # export 'LCOV=lcov 00:02:05.586 --rc lcov_branch_coverage=1 00:02:05.586 --rc lcov_function_coverage=1 00:02:05.586 --rc genhtml_branch_coverage=1 00:02:05.586 --rc genhtml_function_coverage=1 00:02:05.586 --rc genhtml_legend=1 00:02:05.586 --rc geninfo_all_blocks=1 00:02:05.586 --no-external' 00:02:05.586 17:47:59 -- spdk/autotest.sh@81 -- # LCOV='lcov 00:02:05.586 --rc lcov_branch_coverage=1 00:02:05.586 --rc lcov_function_coverage=1 00:02:05.586 --rc genhtml_branch_coverage=1 00:02:05.586 --rc genhtml_function_coverage=1 00:02:05.586 --rc genhtml_legend=1 00:02:05.586 --rc geninfo_all_blocks=1 00:02:05.586 --no-external' 00:02:05.586 17:47:59 -- spdk/autotest.sh@83 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -v 00:02:05.586 lcov: LCOV version 1.14 00:02:05.586 17:47:59 -- spdk/autotest.sh@85 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -c -i -t Baseline -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_base.info 00:02:09.777 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/accel.gcno:no functions found 00:02:09.777 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/accel.gcno 00:02:09.777 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/barrier.gcno:no functions found 00:02:09.777 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/barrier.gcno 00:02:09.777 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/assert.gcno:no functions found 00:02:09.777 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/assert.gcno 00:02:09.777 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/accel_module.gcno:no functions found 00:02:09.777 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/accel_module.gcno 00:02:09.777 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/bdev_zone.gcno:no functions found 00:02:09.777 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/bdev_zone.gcno 00:02:09.777 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/bit_array.gcno:no functions found 00:02:09.777 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/bit_array.gcno 00:02:09.777 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/bit_pool.gcno:no functions found 00:02:09.777 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/bit_pool.gcno 00:02:09.777 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/bdev.gcno:no functions found 00:02:09.777 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/bdev.gcno 00:02:09.777 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/blobfs_bdev.gcno:no functions found 00:02:09.777 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/blobfs_bdev.gcno 00:02:09.777 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/bdev_module.gcno:no functions found 00:02:09.777 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/bdev_module.gcno 00:02:09.777 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/base64.gcno:no functions found 00:02:09.777 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/base64.gcno 00:02:09.777 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/conf.gcno:no functions found 00:02:09.777 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/conf.gcno 00:02:09.777 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/config.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/config.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/blobfs.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/blobfs.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/blob_bdev.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/blob_bdev.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/cpuset.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/cpuset.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/crc16.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/crc16.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/blob.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/blob.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/crc32.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/crc32.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/dif.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/dif.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/endian.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/endian.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/event.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/event.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/dma.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/dma.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/crc64.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/crc64.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/file.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/file.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/env_dpdk.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/env_dpdk.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/ftl.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/ftl.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/fd_group.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/fd_group.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/env.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/env.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/idxd_spec.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/idxd_spec.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/fd.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/fd.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/hexlify.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/hexlify.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/histogram_data.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/histogram_data.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/ioat.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/ioat.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/gpt_spec.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/gpt_spec.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/init.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/init.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/json.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/json.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/iscsi_spec.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/iscsi_spec.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/jsonrpc.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/jsonrpc.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/idxd.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/idxd.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/ioat_spec.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/ioat_spec.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/keyring.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/keyring.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nbd.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nbd.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/mmio.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/mmio.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/keyring_module.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/keyring_module.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/notify.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/notify.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/log.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/log.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/lvol.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/lvol.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/likely.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/likely.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/memory.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/memory.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme_intel.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme_intel.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme_ocssd.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme_ocssd.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme_spec.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme_spec.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvmf.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvmf.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvmf_spec.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvmf_spec.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme_ocssd_spec.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme_ocssd_spec.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme_zns.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvme_zns.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/pci_ids.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/pci_ids.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/pipe.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/pipe.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/opal_spec.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/opal_spec.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvmf_fc_spec.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvmf_fc_spec.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvmf_transport.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvmf_transport.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvmf_cmd.gcno:no functions found 00:02:09.778 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/nvmf_cmd.gcno 00:02:09.778 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/opal.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/opal.gcno 00:02:09.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/queue.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/queue.gcno 00:02:09.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/reduce.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/reduce.gcno 00:02:09.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/rpc.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/rpc.gcno 00:02:09.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/scsi_spec.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/scsi_spec.gcno 00:02:09.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/scsi.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/scsi.gcno 00:02:09.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/sock.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/sock.gcno 00:02:09.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/scheduler.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/scheduler.gcno 00:02:09.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/stdinc.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/stdinc.gcno 00:02:09.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/string.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/string.gcno 00:02:09.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/thread.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/thread.gcno 00:02:09.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/trace.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/trace.gcno 00:02:09.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/trace_parser.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/trace_parser.gcno 00:02:09.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/tree.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/tree.gcno 00:02:09.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/ublk.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/ublk.gcno 00:02:09.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/vfio_user_pci.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/vfio_user_pci.gcno 00:02:09.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/util.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/util.gcno 00:02:09.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/version.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/version.gcno 00:02:09.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/vfio_user_spec.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/vfio_user_spec.gcno 00:02:09.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/uuid.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/uuid.gcno 00:02:09.779 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/vhost.gcno:no functions found 00:02:09.779 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/vhost.gcno 00:02:10.037 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/xor.gcno:no functions found 00:02:10.037 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/xor.gcno 00:02:10.037 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/zipf.gcno:no functions found 00:02:10.037 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/zipf.gcno 00:02:10.037 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/vmd.gcno:no functions found 00:02:10.037 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/cpp_headers/vmd.gcno 00:02:24.956 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/nvme/nvme_stubs.gcno:no functions found 00:02:24.956 geninfo: WARNING: GCOV did not produce any data for /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/nvme/nvme_stubs.gcno 00:02:30.229 17:48:23 -- spdk/autotest.sh@89 -- # timing_enter pre_cleanup 00:02:30.229 17:48:23 -- common/autotest_common.sh@722 -- # xtrace_disable 00:02:30.229 17:48:23 -- common/autotest_common.sh@10 -- # set +x 00:02:30.229 17:48:23 -- spdk/autotest.sh@91 -- # rm -f 00:02:30.229 17:48:23 -- spdk/autotest.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:02:32.767 0000:5e:00.0 (8086 0a54): Already using the nvme driver 00:02:32.767 0000:00:04.7 (8086 2021): Already using the ioatdma driver 00:02:32.767 0000:00:04.6 (8086 2021): Already using the ioatdma driver 00:02:32.767 0000:00:04.5 (8086 2021): Already using the ioatdma driver 00:02:32.767 0000:00:04.4 (8086 2021): Already using the ioatdma driver 00:02:32.767 0000:00:04.3 (8086 2021): Already using the ioatdma driver 00:02:32.767 0000:00:04.2 (8086 2021): Already using the ioatdma driver 00:02:32.767 0000:00:04.1 (8086 2021): Already using the ioatdma driver 00:02:32.767 0000:00:04.0 (8086 2021): Already using the ioatdma driver 00:02:32.767 0000:80:04.7 (8086 2021): Already using the ioatdma driver 00:02:32.767 0000:80:04.6 (8086 2021): Already using the ioatdma driver 00:02:32.767 0000:80:04.5 (8086 2021): Already using the ioatdma driver 00:02:32.767 0000:80:04.4 (8086 2021): Already using the ioatdma driver 00:02:32.767 0000:80:04.3 (8086 2021): Already using the ioatdma driver 00:02:32.767 0000:80:04.2 (8086 2021): Already using the ioatdma driver 00:02:32.767 0000:80:04.1 (8086 2021): Already using the ioatdma driver 00:02:32.767 0000:80:04.0 (8086 2021): Already using the ioatdma driver 00:02:33.026 17:48:26 -- spdk/autotest.sh@96 -- # get_zoned_devs 00:02:33.026 17:48:26 -- common/autotest_common.sh@1669 -- # zoned_devs=() 00:02:33.026 17:48:26 -- common/autotest_common.sh@1669 -- # local -gA zoned_devs 00:02:33.026 17:48:26 -- common/autotest_common.sh@1670 -- # local nvme bdf 00:02:33.027 17:48:26 -- common/autotest_common.sh@1672 -- # for nvme in /sys/block/nvme* 00:02:33.027 17:48:26 -- common/autotest_common.sh@1673 -- # is_block_zoned nvme0n1 00:02:33.027 17:48:26 -- common/autotest_common.sh@1662 -- # local device=nvme0n1 00:02:33.027 17:48:26 -- common/autotest_common.sh@1664 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:02:33.027 17:48:26 -- common/autotest_common.sh@1665 -- # [[ none != none ]] 00:02:33.027 17:48:26 -- spdk/autotest.sh@98 -- # (( 0 > 0 )) 00:02:33.027 17:48:26 -- spdk/autotest.sh@110 -- # for dev in /dev/nvme*n!(*p*) 00:02:33.027 17:48:26 -- spdk/autotest.sh@112 -- # [[ -z '' ]] 00:02:33.027 17:48:26 -- spdk/autotest.sh@113 -- # block_in_use /dev/nvme0n1 00:02:33.027 17:48:26 -- scripts/common.sh@378 -- # local block=/dev/nvme0n1 pt 00:02:33.027 17:48:26 -- scripts/common.sh@387 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py /dev/nvme0n1 00:02:33.027 No valid GPT data, bailing 00:02:33.027 17:48:26 -- scripts/common.sh@391 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:02:33.027 17:48:26 -- scripts/common.sh@391 -- # pt= 00:02:33.027 17:48:26 -- scripts/common.sh@392 -- # return 1 00:02:33.027 17:48:26 -- spdk/autotest.sh@114 -- # dd if=/dev/zero of=/dev/nvme0n1 bs=1M count=1 00:02:33.027 1+0 records in 00:02:33.027 1+0 records out 00:02:33.027 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00466788 s, 225 MB/s 00:02:33.027 17:48:26 -- spdk/autotest.sh@118 -- # sync 00:02:33.027 17:48:26 -- spdk/autotest.sh@120 -- # xtrace_disable_per_cmd reap_spdk_processes 00:02:33.027 17:48:26 -- common/autotest_common.sh@22 -- # eval 'reap_spdk_processes 12> /dev/null' 00:02:33.027 17:48:26 -- common/autotest_common.sh@22 -- # reap_spdk_processes 00:02:38.314 17:48:31 -- spdk/autotest.sh@124 -- # uname -s 00:02:38.314 17:48:31 -- spdk/autotest.sh@124 -- # '[' Linux = Linux ']' 00:02:38.314 17:48:31 -- spdk/autotest.sh@125 -- # run_test setup.sh /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/test-setup.sh 00:02:38.314 17:48:31 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:02:38.314 17:48:31 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:02:38.314 17:48:31 -- common/autotest_common.sh@10 -- # set +x 00:02:38.314 ************************************ 00:02:38.314 START TEST setup.sh 00:02:38.314 ************************************ 00:02:38.314 17:48:31 setup.sh -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/test-setup.sh 00:02:38.314 * Looking for test storage... 00:02:38.314 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup 00:02:38.314 17:48:31 setup.sh -- setup/test-setup.sh@10 -- # uname -s 00:02:38.314 17:48:31 setup.sh -- setup/test-setup.sh@10 -- # [[ Linux == Linux ]] 00:02:38.314 17:48:31 setup.sh -- setup/test-setup.sh@12 -- # run_test acl /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/acl.sh 00:02:38.314 17:48:31 setup.sh -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:02:38.314 17:48:31 setup.sh -- common/autotest_common.sh@1105 -- # xtrace_disable 00:02:38.314 17:48:31 setup.sh -- common/autotest_common.sh@10 -- # set +x 00:02:38.314 ************************************ 00:02:38.314 START TEST acl 00:02:38.314 ************************************ 00:02:38.314 17:48:31 setup.sh.acl -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/acl.sh 00:02:38.314 * Looking for test storage... 00:02:38.314 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup 00:02:38.314 17:48:32 setup.sh.acl -- setup/acl.sh@10 -- # get_zoned_devs 00:02:38.314 17:48:32 setup.sh.acl -- common/autotest_common.sh@1669 -- # zoned_devs=() 00:02:38.314 17:48:32 setup.sh.acl -- common/autotest_common.sh@1669 -- # local -gA zoned_devs 00:02:38.314 17:48:32 setup.sh.acl -- common/autotest_common.sh@1670 -- # local nvme bdf 00:02:38.314 17:48:32 setup.sh.acl -- common/autotest_common.sh@1672 -- # for nvme in /sys/block/nvme* 00:02:38.314 17:48:32 setup.sh.acl -- common/autotest_common.sh@1673 -- # is_block_zoned nvme0n1 00:02:38.314 17:48:32 setup.sh.acl -- common/autotest_common.sh@1662 -- # local device=nvme0n1 00:02:38.314 17:48:32 setup.sh.acl -- common/autotest_common.sh@1664 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:02:38.314 17:48:32 setup.sh.acl -- common/autotest_common.sh@1665 -- # [[ none != none ]] 00:02:38.314 17:48:32 setup.sh.acl -- setup/acl.sh@12 -- # devs=() 00:02:38.314 17:48:32 setup.sh.acl -- setup/acl.sh@12 -- # declare -a devs 00:02:38.314 17:48:32 setup.sh.acl -- setup/acl.sh@13 -- # drivers=() 00:02:38.314 17:48:32 setup.sh.acl -- setup/acl.sh@13 -- # declare -A drivers 00:02:38.314 17:48:32 setup.sh.acl -- setup/acl.sh@51 -- # setup reset 00:02:38.314 17:48:32 setup.sh.acl -- setup/common.sh@9 -- # [[ reset == output ]] 00:02:38.314 17:48:32 setup.sh.acl -- setup/common.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:02:41.602 17:48:34 setup.sh.acl -- setup/acl.sh@52 -- # collect_setup_devs 00:02:41.602 17:48:34 setup.sh.acl -- setup/acl.sh@16 -- # local dev driver 00:02:41.602 17:48:34 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:41.602 17:48:34 setup.sh.acl -- setup/acl.sh@15 -- # setup output status 00:02:41.602 17:48:34 setup.sh.acl -- setup/common.sh@9 -- # [[ output == output ]] 00:02:41.602 17:48:34 setup.sh.acl -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh status 00:02:43.504 Hugepages 00:02:43.504 node hugesize free / total 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 1048576kB == *:*:*.* ]] 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # continue 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 2048kB == *:*:*.* ]] 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # continue 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 1048576kB == *:*:*.* ]] 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # continue 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.504 00:02:43.504 Type BDF Vendor Device NUMA Driver Device Block devices 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 2048kB == *:*:*.* ]] 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # continue 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:00:04.0 == *:*:*.* ]] 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:00:04.1 == *:*:*.* ]] 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:00:04.2 == *:*:*.* ]] 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:02:43.504 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:02:43.505 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.505 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:00:04.3 == *:*:*.* ]] 00:02:43.505 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:02:43.505 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:02:43.505 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.505 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:00:04.4 == *:*:*.* ]] 00:02:43.505 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:02:43.505 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:02:43.505 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.505 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:00:04.5 == *:*:*.* ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:00:04.6 == *:*:*.* ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:00:04.7 == *:*:*.* ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:5e:00.0 == *:*:*.* ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # [[ nvme == nvme ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@21 -- # [[ '' == *\0\0\0\0\:\5\e\:\0\0\.\0* ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@22 -- # devs+=("$dev") 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@22 -- # drivers["$dev"]=nvme 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:80:04.0 == *:*:*.* ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:80:04.1 == *:*:*.* ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:80:04.2 == *:*:*.* ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:80:04.3 == *:*:*.* ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:80:04.4 == *:*:*.* ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:80:04.5 == *:*:*.* ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:80:04.6 == *:*:*.* ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@19 -- # [[ 0000:80:04.7 == *:*:*.* ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # [[ ioatdma == nvme ]] 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@20 -- # continue 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@18 -- # read -r _ dev _ _ _ driver _ 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@24 -- # (( 1 > 0 )) 00:02:43.763 17:48:37 setup.sh.acl -- setup/acl.sh@54 -- # run_test denied denied 00:02:43.763 17:48:37 setup.sh.acl -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:02:43.763 17:48:37 setup.sh.acl -- common/autotest_common.sh@1105 -- # xtrace_disable 00:02:43.763 17:48:37 setup.sh.acl -- common/autotest_common.sh@10 -- # set +x 00:02:43.763 ************************************ 00:02:43.763 START TEST denied 00:02:43.763 ************************************ 00:02:43.763 17:48:37 setup.sh.acl.denied -- common/autotest_common.sh@1123 -- # denied 00:02:43.763 17:48:37 setup.sh.acl.denied -- setup/acl.sh@38 -- # PCI_BLOCKED=' 0000:5e:00.0' 00:02:43.763 17:48:37 setup.sh.acl.denied -- setup/acl.sh@38 -- # setup output config 00:02:43.763 17:48:37 setup.sh.acl.denied -- setup/acl.sh@39 -- # grep 'Skipping denied controller at 0000:5e:00.0' 00:02:43.763 17:48:37 setup.sh.acl.denied -- setup/common.sh@9 -- # [[ output == output ]] 00:02:43.763 17:48:37 setup.sh.acl.denied -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh config 00:02:46.291 0000:5e:00.0 (8086 0a54): Skipping denied controller at 0000:5e:00.0 00:02:46.291 17:48:39 setup.sh.acl.denied -- setup/acl.sh@40 -- # verify 0000:5e:00.0 00:02:46.291 17:48:39 setup.sh.acl.denied -- setup/acl.sh@28 -- # local dev driver 00:02:46.291 17:48:39 setup.sh.acl.denied -- setup/acl.sh@30 -- # for dev in "$@" 00:02:46.291 17:48:39 setup.sh.acl.denied -- setup/acl.sh@31 -- # [[ -e /sys/bus/pci/devices/0000:5e:00.0 ]] 00:02:46.291 17:48:39 setup.sh.acl.denied -- setup/acl.sh@32 -- # readlink -f /sys/bus/pci/devices/0000:5e:00.0/driver 00:02:46.291 17:48:39 setup.sh.acl.denied -- setup/acl.sh@32 -- # driver=/sys/bus/pci/drivers/nvme 00:02:46.291 17:48:39 setup.sh.acl.denied -- setup/acl.sh@33 -- # [[ nvme == \n\v\m\e ]] 00:02:46.291 17:48:39 setup.sh.acl.denied -- setup/acl.sh@41 -- # setup reset 00:02:46.291 17:48:39 setup.sh.acl.denied -- setup/common.sh@9 -- # [[ reset == output ]] 00:02:46.291 17:48:39 setup.sh.acl.denied -- setup/common.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:02:50.512 00:02:50.512 real 0m6.065s 00:02:50.512 user 0m1.831s 00:02:50.512 sys 0m3.416s 00:02:50.512 17:48:43 setup.sh.acl.denied -- common/autotest_common.sh@1124 -- # xtrace_disable 00:02:50.512 17:48:43 setup.sh.acl.denied -- common/autotest_common.sh@10 -- # set +x 00:02:50.512 ************************************ 00:02:50.512 END TEST denied 00:02:50.512 ************************************ 00:02:50.512 17:48:43 setup.sh.acl -- common/autotest_common.sh@1142 -- # return 0 00:02:50.512 17:48:43 setup.sh.acl -- setup/acl.sh@55 -- # run_test allowed allowed 00:02:50.512 17:48:43 setup.sh.acl -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:02:50.512 17:48:43 setup.sh.acl -- common/autotest_common.sh@1105 -- # xtrace_disable 00:02:50.512 17:48:43 setup.sh.acl -- common/autotest_common.sh@10 -- # set +x 00:02:50.512 ************************************ 00:02:50.512 START TEST allowed 00:02:50.512 ************************************ 00:02:50.512 17:48:43 setup.sh.acl.allowed -- common/autotest_common.sh@1123 -- # allowed 00:02:50.512 17:48:43 setup.sh.acl.allowed -- setup/acl.sh@45 -- # PCI_ALLOWED=0000:5e:00.0 00:02:50.512 17:48:43 setup.sh.acl.allowed -- setup/acl.sh@45 -- # setup output config 00:02:50.512 17:48:43 setup.sh.acl.allowed -- setup/acl.sh@46 -- # grep -E '0000:5e:00.0 .*: nvme -> .*' 00:02:50.512 17:48:43 setup.sh.acl.allowed -- setup/common.sh@9 -- # [[ output == output ]] 00:02:50.512 17:48:43 setup.sh.acl.allowed -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh config 00:02:53.815 0000:5e:00.0 (8086 0a54): nvme -> vfio-pci 00:02:53.815 17:48:46 setup.sh.acl.allowed -- setup/acl.sh@47 -- # verify 00:02:53.815 17:48:46 setup.sh.acl.allowed -- setup/acl.sh@28 -- # local dev driver 00:02:53.815 17:48:46 setup.sh.acl.allowed -- setup/acl.sh@48 -- # setup reset 00:02:53.815 17:48:46 setup.sh.acl.allowed -- setup/common.sh@9 -- # [[ reset == output ]] 00:02:53.815 17:48:46 setup.sh.acl.allowed -- setup/common.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:02:56.347 00:02:56.347 real 0m6.093s 00:02:56.347 user 0m1.797s 00:02:56.347 sys 0m3.302s 00:02:56.347 17:48:49 setup.sh.acl.allowed -- common/autotest_common.sh@1124 -- # xtrace_disable 00:02:56.347 17:48:49 setup.sh.acl.allowed -- common/autotest_common.sh@10 -- # set +x 00:02:56.347 ************************************ 00:02:56.347 END TEST allowed 00:02:56.347 ************************************ 00:02:56.347 17:48:49 setup.sh.acl -- common/autotest_common.sh@1142 -- # return 0 00:02:56.347 00:02:56.347 real 0m17.718s 00:02:56.347 user 0m5.651s 00:02:56.347 sys 0m10.285s 00:02:56.347 17:48:49 setup.sh.acl -- common/autotest_common.sh@1124 -- # xtrace_disable 00:02:56.347 17:48:49 setup.sh.acl -- common/autotest_common.sh@10 -- # set +x 00:02:56.347 ************************************ 00:02:56.347 END TEST acl 00:02:56.347 ************************************ 00:02:56.347 17:48:49 setup.sh -- common/autotest_common.sh@1142 -- # return 0 00:02:56.347 17:48:49 setup.sh -- setup/test-setup.sh@13 -- # run_test hugepages /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/hugepages.sh 00:02:56.347 17:48:49 setup.sh -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:02:56.347 17:48:49 setup.sh -- common/autotest_common.sh@1105 -- # xtrace_disable 00:02:56.347 17:48:49 setup.sh -- common/autotest_common.sh@10 -- # set +x 00:02:56.347 ************************************ 00:02:56.347 START TEST hugepages 00:02:56.347 ************************************ 00:02:56.347 17:48:49 setup.sh.hugepages -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/hugepages.sh 00:02:56.347 * Looking for test storage... 00:02:56.347 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@10 -- # nodes_sys=() 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@10 -- # declare -a nodes_sys 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@12 -- # declare -i default_hugepages=0 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@13 -- # declare -i no_nodes=0 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@14 -- # declare -i nr_hugepages=0 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@16 -- # get_meminfo Hugepagesize 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/common.sh@17 -- # local get=Hugepagesize 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/common.sh@18 -- # local node= 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/common.sh@19 -- # local var val 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/common.sh@20 -- # local mem_f mem 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/common.sh@25 -- # [[ -n '' ]] 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/common.sh@28 -- # mapfile -t mem 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 173372816 kB' 'MemAvailable: 176244008 kB' 'Buffers: 3896 kB' 'Cached: 10112276 kB' 'SwapCached: 0 kB' 'Active: 7124320 kB' 'Inactive: 3507524 kB' 'Active(anon): 6732312 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 518988 kB' 'Mapped: 171348 kB' 'Shmem: 6216640 kB' 'KReclaimable: 232220 kB' 'Slab: 813176 kB' 'SReclaimable: 232220 kB' 'SUnreclaim: 580956 kB' 'KernelStack: 20480 kB' 'PageTables: 8728 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 101982028 kB' 'Committed_AS: 8268296 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315404 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 2048' 'HugePages_Free: 2048' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 4194304 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.347 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.348 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # continue 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # IFS=': ' 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@31 -- # read -r var val _ 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@32 -- # [[ Hugepagesize == \H\u\g\e\p\a\g\e\s\i\z\e ]] 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@33 -- # echo 2048 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/common.sh@33 -- # return 0 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@16 -- # default_hugepages=2048 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@17 -- # default_huge_nr=/sys/kernel/mm/hugepages/hugepages-2048kB/nr_hugepages 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@18 -- # global_huge_nr=/proc/sys/vm/nr_hugepages 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@21 -- # unset -v HUGE_EVEN_ALLOC 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@22 -- # unset -v HUGEMEM 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@23 -- # unset -v HUGENODE 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@24 -- # unset -v NRHUGE 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@207 -- # get_nodes 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@27 -- # local node 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=2048 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=0 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@32 -- # no_nodes=2 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@208 -- # clear_hp 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@37 -- # local node hp 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@39 -- # for node in "${!nodes_sys[@]}" 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@41 -- # echo 0 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@41 -- # echo 0 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@39 -- # for node in "${!nodes_sys[@]}" 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@41 -- # echo 0 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@41 -- # echo 0 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@45 -- # export CLEAR_HUGE=yes 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@45 -- # CLEAR_HUGE=yes 00:02:56.349 17:48:49 setup.sh.hugepages -- setup/hugepages.sh@210 -- # run_test default_setup default_setup 00:02:56.349 17:48:49 setup.sh.hugepages -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:02:56.349 17:48:49 setup.sh.hugepages -- common/autotest_common.sh@1105 -- # xtrace_disable 00:02:56.349 17:48:49 setup.sh.hugepages -- common/autotest_common.sh@10 -- # set +x 00:02:56.349 ************************************ 00:02:56.349 START TEST default_setup 00:02:56.349 ************************************ 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- common/autotest_common.sh@1123 -- # default_setup 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/hugepages.sh@136 -- # get_test_nr_hugepages 2097152 0 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/hugepages.sh@49 -- # local size=2097152 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/hugepages.sh@50 -- # (( 2 > 1 )) 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/hugepages.sh@51 -- # shift 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/hugepages.sh@52 -- # node_ids=('0') 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/hugepages.sh@52 -- # local node_ids 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/hugepages.sh@57 -- # nr_hugepages=1024 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 0 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/hugepages.sh@62 -- # user_nodes=('0') 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/hugepages.sh@62 -- # local user_nodes 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/hugepages.sh@64 -- # local _nr_hugepages=1024 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/hugepages.sh@67 -- # nodes_test=() 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/hugepages.sh@67 -- # local -g nodes_test 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/hugepages.sh@69 -- # (( 1 > 0 )) 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/hugepages.sh@70 -- # for _no_nodes in "${user_nodes[@]}" 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/hugepages.sh@71 -- # nodes_test[_no_nodes]=1024 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/hugepages.sh@73 -- # return 0 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/hugepages.sh@137 -- # setup output 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/common.sh@9 -- # [[ output == output ]] 00:02:56.349 17:48:49 setup.sh.hugepages.default_setup -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:02:58.881 0000:00:04.7 (8086 2021): ioatdma -> vfio-pci 00:02:58.881 0000:00:04.6 (8086 2021): ioatdma -> vfio-pci 00:02:58.881 0000:00:04.5 (8086 2021): ioatdma -> vfio-pci 00:02:58.881 0000:00:04.4 (8086 2021): ioatdma -> vfio-pci 00:02:58.881 0000:00:04.3 (8086 2021): ioatdma -> vfio-pci 00:02:58.881 0000:00:04.2 (8086 2021): ioatdma -> vfio-pci 00:02:59.140 0000:00:04.1 (8086 2021): ioatdma -> vfio-pci 00:02:59.140 0000:00:04.0 (8086 2021): ioatdma -> vfio-pci 00:02:59.140 0000:80:04.7 (8086 2021): ioatdma -> vfio-pci 00:02:59.140 0000:80:04.6 (8086 2021): ioatdma -> vfio-pci 00:02:59.140 0000:80:04.5 (8086 2021): ioatdma -> vfio-pci 00:02:59.140 0000:80:04.4 (8086 2021): ioatdma -> vfio-pci 00:02:59.140 0000:80:04.3 (8086 2021): ioatdma -> vfio-pci 00:02:59.140 0000:80:04.2 (8086 2021): ioatdma -> vfio-pci 00:02:59.140 0000:80:04.1 (8086 2021): ioatdma -> vfio-pci 00:02:59.140 0000:80:04.0 (8086 2021): ioatdma -> vfio-pci 00:03:00.082 0000:5e:00.0 (8086 0a54): nvme -> vfio-pci 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@138 -- # verify_nr_hugepages 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@89 -- # local node 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@90 -- # local sorted_t 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@91 -- # local sorted_s 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@92 -- # local surp 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@93 -- # local resv 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@94 -- # local anon 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@17 -- # local get=AnonHugePages 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@18 -- # local node= 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@19 -- # local var val 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@20 -- # local mem_f mem 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@28 -- # mapfile -t mem 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175563064 kB' 'MemAvailable: 178434228 kB' 'Buffers: 3896 kB' 'Cached: 10112380 kB' 'SwapCached: 0 kB' 'Active: 7139076 kB' 'Inactive: 3507524 kB' 'Active(anon): 6747068 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 533492 kB' 'Mapped: 171116 kB' 'Shmem: 6216744 kB' 'KReclaimable: 232164 kB' 'Slab: 811180 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 579016 kB' 'KernelStack: 20512 kB' 'PageTables: 8724 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103030604 kB' 'Committed_AS: 8287512 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315676 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.082 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.083 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@33 -- # echo 0 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@33 -- # return 0 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@97 -- # anon=0 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@17 -- # local get=HugePages_Surp 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@18 -- # local node= 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@19 -- # local var val 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@20 -- # local mem_f mem 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@28 -- # mapfile -t mem 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175563392 kB' 'MemAvailable: 178434556 kB' 'Buffers: 3896 kB' 'Cached: 10112384 kB' 'SwapCached: 0 kB' 'Active: 7140656 kB' 'Inactive: 3507524 kB' 'Active(anon): 6748648 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 535056 kB' 'Mapped: 171080 kB' 'Shmem: 6216748 kB' 'KReclaimable: 232164 kB' 'Slab: 811212 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 579048 kB' 'KernelStack: 20768 kB' 'PageTables: 9432 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103030604 kB' 'Committed_AS: 8287532 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315740 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.084 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@33 -- # echo 0 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@33 -- # return 0 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@99 -- # surp=0 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@18 -- # local node= 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@19 -- # local var val 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@20 -- # local mem_f mem 00:03:00.085 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@28 -- # mapfile -t mem 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175566316 kB' 'MemAvailable: 178437480 kB' 'Buffers: 3896 kB' 'Cached: 10112396 kB' 'SwapCached: 0 kB' 'Active: 7142772 kB' 'Inactive: 3507524 kB' 'Active(anon): 6750764 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 537192 kB' 'Mapped: 171584 kB' 'Shmem: 6216760 kB' 'KReclaimable: 232164 kB' 'Slab: 811300 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 579136 kB' 'KernelStack: 21136 kB' 'PageTables: 10524 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103030604 kB' 'Committed_AS: 8289964 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315756 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.086 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@33 -- # echo 0 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@33 -- # return 0 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@100 -- # resv=0 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@102 -- # echo nr_hugepages=1024 00:03:00.087 nr_hugepages=1024 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:03:00.087 resv_hugepages=0 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:03:00.087 surplus_hugepages=0 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:03:00.087 anon_hugepages=0 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@107 -- # (( 1024 == nr_hugepages + surp + resv )) 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@109 -- # (( 1024 == nr_hugepages )) 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@17 -- # local get=HugePages_Total 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@18 -- # local node= 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@19 -- # local var val 00:03:00.087 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@20 -- # local mem_f mem 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@28 -- # mapfile -t mem 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175560304 kB' 'MemAvailable: 178431468 kB' 'Buffers: 3896 kB' 'Cached: 10112420 kB' 'SwapCached: 0 kB' 'Active: 7146648 kB' 'Inactive: 3507524 kB' 'Active(anon): 6754640 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 540988 kB' 'Mapped: 171984 kB' 'Shmem: 6216784 kB' 'KReclaimable: 232164 kB' 'Slab: 811108 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 578944 kB' 'KernelStack: 21008 kB' 'PageTables: 10012 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103030604 kB' 'Committed_AS: 8293696 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315728 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.088 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@33 -- # echo 1024 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@33 -- # return 0 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@110 -- # (( 1024 == nr_hugepages + surp + resv )) 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@112 -- # get_nodes 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@27 -- # local node 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=1024 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=0 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@32 -- # no_nodes=2 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@17 -- # local get=HugePages_Surp 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@18 -- # local node=0 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@19 -- # local var val 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@20 -- # local mem_f mem 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@28 -- # mapfile -t mem 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 97662684 kB' 'MemFree: 85807424 kB' 'MemUsed: 11855260 kB' 'SwapCached: 0 kB' 'Active: 4909936 kB' 'Inactive: 3335920 kB' 'Active(anon): 4752396 kB' 'Inactive(anon): 0 kB' 'Active(file): 157540 kB' 'Inactive(file): 3335920 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8121488 kB' 'Mapped: 89772 kB' 'AnonPages: 127412 kB' 'Shmem: 4628028 kB' 'KernelStack: 12232 kB' 'PageTables: 6536 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 124964 kB' 'Slab: 393716 kB' 'SReclaimable: 124964 kB' 'SUnreclaim: 268752 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Surp: 0' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.089 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # continue 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # IFS=': ' 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@31 -- # read -r var val _ 00:03:00.090 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:00.091 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@33 -- # echo 0 00:03:00.091 17:48:53 setup.sh.hugepages.default_setup -- setup/common.sh@33 -- # return 0 00:03:00.091 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:03:00.091 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:03:00.091 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:03:00.091 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:03:00.091 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@128 -- # echo 'node0=1024 expecting 1024' 00:03:00.091 node0=1024 expecting 1024 00:03:00.091 17:48:53 setup.sh.hugepages.default_setup -- setup/hugepages.sh@130 -- # [[ 1024 == \1\0\2\4 ]] 00:03:00.091 00:03:00.091 real 0m3.912s 00:03:00.091 user 0m1.293s 00:03:00.091 sys 0m1.865s 00:03:00.091 17:48:53 setup.sh.hugepages.default_setup -- common/autotest_common.sh@1124 -- # xtrace_disable 00:03:00.091 17:48:53 setup.sh.hugepages.default_setup -- common/autotest_common.sh@10 -- # set +x 00:03:00.091 ************************************ 00:03:00.091 END TEST default_setup 00:03:00.091 ************************************ 00:03:00.350 17:48:53 setup.sh.hugepages -- common/autotest_common.sh@1142 -- # return 0 00:03:00.350 17:48:53 setup.sh.hugepages -- setup/hugepages.sh@211 -- # run_test per_node_1G_alloc per_node_1G_alloc 00:03:00.350 17:48:53 setup.sh.hugepages -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:03:00.350 17:48:53 setup.sh.hugepages -- common/autotest_common.sh@1105 -- # xtrace_disable 00:03:00.350 17:48:53 setup.sh.hugepages -- common/autotest_common.sh@10 -- # set +x 00:03:00.350 ************************************ 00:03:00.350 START TEST per_node_1G_alloc 00:03:00.350 ************************************ 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- common/autotest_common.sh@1123 -- # per_node_1G_alloc 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@143 -- # local IFS=, 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@145 -- # get_test_nr_hugepages 1048576 0 1 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@49 -- # local size=1048576 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@50 -- # (( 3 > 1 )) 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@51 -- # shift 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@52 -- # node_ids=('0' '1') 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@52 -- # local node_ids 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@57 -- # nr_hugepages=512 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 0 1 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@62 -- # user_nodes=('0' '1') 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@62 -- # local user_nodes 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@64 -- # local _nr_hugepages=512 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@67 -- # nodes_test=() 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@67 -- # local -g nodes_test 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@69 -- # (( 2 > 0 )) 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@70 -- # for _no_nodes in "${user_nodes[@]}" 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@71 -- # nodes_test[_no_nodes]=512 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@70 -- # for _no_nodes in "${user_nodes[@]}" 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@71 -- # nodes_test[_no_nodes]=512 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@73 -- # return 0 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@146 -- # NRHUGE=512 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@146 -- # HUGENODE=0,1 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@146 -- # setup output 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@9 -- # [[ output == output ]] 00:03:00.350 17:48:53 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:03:02.880 0000:00:04.7 (8086 2021): Already using the vfio-pci driver 00:03:02.880 0000:5e:00.0 (8086 0a54): Already using the vfio-pci driver 00:03:02.880 0000:00:04.6 (8086 2021): Already using the vfio-pci driver 00:03:02.880 0000:00:04.5 (8086 2021): Already using the vfio-pci driver 00:03:02.880 0000:00:04.4 (8086 2021): Already using the vfio-pci driver 00:03:02.880 0000:00:04.3 (8086 2021): Already using the vfio-pci driver 00:03:02.880 0000:00:04.2 (8086 2021): Already using the vfio-pci driver 00:03:02.880 0000:00:04.1 (8086 2021): Already using the vfio-pci driver 00:03:02.880 0000:00:04.0 (8086 2021): Already using the vfio-pci driver 00:03:02.880 0000:80:04.7 (8086 2021): Already using the vfio-pci driver 00:03:02.880 0000:80:04.6 (8086 2021): Already using the vfio-pci driver 00:03:02.880 0000:80:04.5 (8086 2021): Already using the vfio-pci driver 00:03:02.880 0000:80:04.4 (8086 2021): Already using the vfio-pci driver 00:03:02.880 0000:80:04.3 (8086 2021): Already using the vfio-pci driver 00:03:02.880 0000:80:04.2 (8086 2021): Already using the vfio-pci driver 00:03:02.880 0000:80:04.1 (8086 2021): Already using the vfio-pci driver 00:03:02.880 0000:80:04.0 (8086 2021): Already using the vfio-pci driver 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@147 -- # nr_hugepages=1024 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@147 -- # verify_nr_hugepages 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@89 -- # local node 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@90 -- # local sorted_t 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@91 -- # local sorted_s 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@92 -- # local surp 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@93 -- # local resv 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@94 -- # local anon 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@17 -- # local get=AnonHugePages 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@18 -- # local node= 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@19 -- # local var val 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175571064 kB' 'MemAvailable: 178442228 kB' 'Buffers: 3896 kB' 'Cached: 10112520 kB' 'SwapCached: 0 kB' 'Active: 7139744 kB' 'Inactive: 3507524 kB' 'Active(anon): 6747736 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 534056 kB' 'Mapped: 171096 kB' 'Shmem: 6216884 kB' 'KReclaimable: 232164 kB' 'Slab: 811660 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 579496 kB' 'KernelStack: 20544 kB' 'PageTables: 8736 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103030604 kB' 'Committed_AS: 8285552 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315612 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.146 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # echo 0 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # return 0 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@97 -- # anon=0 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@18 -- # local node= 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@19 -- # local var val 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175572836 kB' 'MemAvailable: 178444000 kB' 'Buffers: 3896 kB' 'Cached: 10112524 kB' 'SwapCached: 0 kB' 'Active: 7139980 kB' 'Inactive: 3507524 kB' 'Active(anon): 6747972 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 534320 kB' 'Mapped: 171096 kB' 'Shmem: 6216888 kB' 'KReclaimable: 232164 kB' 'Slab: 811684 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 579520 kB' 'KernelStack: 20544 kB' 'PageTables: 8732 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103030604 kB' 'Committed_AS: 8285572 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315580 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.147 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # echo 0 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # return 0 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@99 -- # surp=0 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@18 -- # local node= 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@19 -- # local var val 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175572860 kB' 'MemAvailable: 178444024 kB' 'Buffers: 3896 kB' 'Cached: 10112540 kB' 'SwapCached: 0 kB' 'Active: 7139992 kB' 'Inactive: 3507524 kB' 'Active(anon): 6747984 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 534320 kB' 'Mapped: 171096 kB' 'Shmem: 6216904 kB' 'KReclaimable: 232164 kB' 'Slab: 811684 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 579520 kB' 'KernelStack: 20544 kB' 'PageTables: 8732 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103030604 kB' 'Committed_AS: 8285596 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315580 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.148 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.149 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # echo 0 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # return 0 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@100 -- # resv=0 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@102 -- # echo nr_hugepages=1024 00:03:03.150 nr_hugepages=1024 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:03:03.150 resv_hugepages=0 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:03:03.150 surplus_hugepages=0 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:03:03.150 anon_hugepages=0 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@107 -- # (( 1024 == nr_hugepages + surp + resv )) 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@109 -- # (( 1024 == nr_hugepages )) 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@17 -- # local get=HugePages_Total 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@18 -- # local node= 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@19 -- # local var val 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175572920 kB' 'MemAvailable: 178444084 kB' 'Buffers: 3896 kB' 'Cached: 10112540 kB' 'SwapCached: 0 kB' 'Active: 7139636 kB' 'Inactive: 3507524 kB' 'Active(anon): 6747628 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 533964 kB' 'Mapped: 171096 kB' 'Shmem: 6216904 kB' 'KReclaimable: 232164 kB' 'Slab: 811684 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 579520 kB' 'KernelStack: 20528 kB' 'PageTables: 8680 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103030604 kB' 'Committed_AS: 8285616 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315580 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.150 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # echo 1024 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # return 0 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@110 -- # (( 1024 == nr_hugepages + surp + resv )) 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@112 -- # get_nodes 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@27 -- # local node 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=512 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=512 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@32 -- # no_nodes=2 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@18 -- # local node=0 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@19 -- # local var val 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 97662684 kB' 'MemFree: 86868644 kB' 'MemUsed: 10794040 kB' 'SwapCached: 0 kB' 'Active: 4909444 kB' 'Inactive: 3335920 kB' 'Active(anon): 4751904 kB' 'Inactive(anon): 0 kB' 'Active(file): 157540 kB' 'Inactive(file): 3335920 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8121520 kB' 'Mapped: 89720 kB' 'AnonPages: 127000 kB' 'Shmem: 4628060 kB' 'KernelStack: 11464 kB' 'PageTables: 3948 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 124964 kB' 'Slab: 394364 kB' 'SReclaimable: 124964 kB' 'SUnreclaim: 269400 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 512' 'HugePages_Free: 512' 'HugePages_Surp: 0' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.151 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # echo 0 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # return 0 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 1 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@18 -- # local node=1 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@19 -- # local var val 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node1/meminfo ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node1/meminfo 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 93718468 kB' 'MemFree: 88705224 kB' 'MemUsed: 5013244 kB' 'SwapCached: 0 kB' 'Active: 2230644 kB' 'Inactive: 171604 kB' 'Active(anon): 1996176 kB' 'Inactive(anon): 0 kB' 'Active(file): 234468 kB' 'Inactive(file): 171604 kB' 'Unevictable: 0 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 1994984 kB' 'Mapped: 81376 kB' 'AnonPages: 407328 kB' 'Shmem: 1588912 kB' 'KernelStack: 9080 kB' 'PageTables: 4784 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 107200 kB' 'Slab: 417320 kB' 'SReclaimable: 107200 kB' 'SUnreclaim: 310120 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 512' 'HugePages_Free: 512' 'HugePages_Surp: 0' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.152 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # continue 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # echo 0 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/common.sh@33 -- # return 0 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@128 -- # echo 'node0=512 expecting 512' 00:03:03.153 node0=512 expecting 512 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@128 -- # echo 'node1=512 expecting 512' 00:03:03.153 node1=512 expecting 512 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- setup/hugepages.sh@130 -- # [[ 512 == \5\1\2 ]] 00:03:03.153 00:03:03.153 real 0m2.969s 00:03:03.153 user 0m1.197s 00:03:03.153 sys 0m1.833s 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:03:03.153 17:48:56 setup.sh.hugepages.per_node_1G_alloc -- common/autotest_common.sh@10 -- # set +x 00:03:03.153 ************************************ 00:03:03.153 END TEST per_node_1G_alloc 00:03:03.153 ************************************ 00:03:03.411 17:48:56 setup.sh.hugepages -- common/autotest_common.sh@1142 -- # return 0 00:03:03.411 17:48:56 setup.sh.hugepages -- setup/hugepages.sh@212 -- # run_test even_2G_alloc even_2G_alloc 00:03:03.411 17:48:56 setup.sh.hugepages -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:03:03.411 17:48:56 setup.sh.hugepages -- common/autotest_common.sh@1105 -- # xtrace_disable 00:03:03.411 17:48:56 setup.sh.hugepages -- common/autotest_common.sh@10 -- # set +x 00:03:03.411 ************************************ 00:03:03.412 START TEST even_2G_alloc 00:03:03.412 ************************************ 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- common/autotest_common.sh@1123 -- # even_2G_alloc 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@152 -- # get_test_nr_hugepages 2097152 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@49 -- # local size=2097152 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@50 -- # (( 1 > 1 )) 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@57 -- # nr_hugepages=1024 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@62 -- # user_nodes=() 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@62 -- # local user_nodes 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@64 -- # local _nr_hugepages=1024 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@67 -- # nodes_test=() 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@67 -- # local -g nodes_test 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@69 -- # (( 0 > 0 )) 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@74 -- # (( 0 > 0 )) 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@82 -- # nodes_test[_no_nodes - 1]=512 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@83 -- # : 512 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@84 -- # : 1 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@82 -- # nodes_test[_no_nodes - 1]=512 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@83 -- # : 0 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@84 -- # : 0 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@153 -- # NRHUGE=1024 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@153 -- # HUGE_EVEN_ALLOC=yes 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@153 -- # setup output 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@9 -- # [[ output == output ]] 00:03:03.412 17:48:56 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:03:05.939 0000:00:04.7 (8086 2021): Already using the vfio-pci driver 00:03:05.939 0000:5e:00.0 (8086 0a54): Already using the vfio-pci driver 00:03:05.939 0000:00:04.6 (8086 2021): Already using the vfio-pci driver 00:03:05.939 0000:00:04.5 (8086 2021): Already using the vfio-pci driver 00:03:05.939 0000:00:04.4 (8086 2021): Already using the vfio-pci driver 00:03:05.939 0000:00:04.3 (8086 2021): Already using the vfio-pci driver 00:03:05.939 0000:00:04.2 (8086 2021): Already using the vfio-pci driver 00:03:05.939 0000:00:04.1 (8086 2021): Already using the vfio-pci driver 00:03:05.939 0000:00:04.0 (8086 2021): Already using the vfio-pci driver 00:03:05.939 0000:80:04.7 (8086 2021): Already using the vfio-pci driver 00:03:05.939 0000:80:04.6 (8086 2021): Already using the vfio-pci driver 00:03:05.939 0000:80:04.5 (8086 2021): Already using the vfio-pci driver 00:03:05.939 0000:80:04.4 (8086 2021): Already using the vfio-pci driver 00:03:05.939 0000:80:04.3 (8086 2021): Already using the vfio-pci driver 00:03:05.939 0000:80:04.2 (8086 2021): Already using the vfio-pci driver 00:03:05.939 0000:80:04.1 (8086 2021): Already using the vfio-pci driver 00:03:05.939 0000:80:04.0 (8086 2021): Already using the vfio-pci driver 00:03:06.204 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@154 -- # verify_nr_hugepages 00:03:06.204 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@89 -- # local node 00:03:06.204 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@90 -- # local sorted_t 00:03:06.204 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@91 -- # local sorted_s 00:03:06.204 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@92 -- # local surp 00:03:06.204 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@93 -- # local resv 00:03:06.204 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@94 -- # local anon 00:03:06.204 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:03:06.204 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:03:06.204 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@17 -- # local get=AnonHugePages 00:03:06.204 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@18 -- # local node= 00:03:06.204 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@19 -- # local var val 00:03:06.204 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:06.204 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:06.204 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:06.204 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175631872 kB' 'MemAvailable: 178503036 kB' 'Buffers: 3896 kB' 'Cached: 10112680 kB' 'SwapCached: 0 kB' 'Active: 7138268 kB' 'Inactive: 3507524 kB' 'Active(anon): 6746260 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 532468 kB' 'Mapped: 170088 kB' 'Shmem: 6217044 kB' 'KReclaimable: 232164 kB' 'Slab: 811100 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 578936 kB' 'KernelStack: 20608 kB' 'PageTables: 8900 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103030604 kB' 'Committed_AS: 8274684 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315548 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.205 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # echo 0 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # return 0 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@97 -- # anon=0 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@18 -- # local node= 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@19 -- # local var val 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.206 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175632328 kB' 'MemAvailable: 178503492 kB' 'Buffers: 3896 kB' 'Cached: 10112684 kB' 'SwapCached: 0 kB' 'Active: 7138532 kB' 'Inactive: 3507524 kB' 'Active(anon): 6746524 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 532732 kB' 'Mapped: 170052 kB' 'Shmem: 6217048 kB' 'KReclaimable: 232164 kB' 'Slab: 811128 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 578964 kB' 'KernelStack: 20624 kB' 'PageTables: 8960 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103030604 kB' 'Committed_AS: 8274700 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315500 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.207 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.208 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # echo 0 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # return 0 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@99 -- # surp=0 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@18 -- # local node= 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@19 -- # local var val 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175632584 kB' 'MemAvailable: 178503748 kB' 'Buffers: 3896 kB' 'Cached: 10112684 kB' 'SwapCached: 0 kB' 'Active: 7138568 kB' 'Inactive: 3507524 kB' 'Active(anon): 6746560 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 532768 kB' 'Mapped: 170052 kB' 'Shmem: 6217048 kB' 'KReclaimable: 232164 kB' 'Slab: 811128 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 578964 kB' 'KernelStack: 20640 kB' 'PageTables: 9012 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103030604 kB' 'Committed_AS: 8274720 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315500 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.209 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.210 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # echo 0 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # return 0 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@100 -- # resv=0 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@102 -- # echo nr_hugepages=1024 00:03:06.211 nr_hugepages=1024 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:03:06.211 resv_hugepages=0 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:03:06.211 surplus_hugepages=0 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:03:06.211 anon_hugepages=0 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@107 -- # (( 1024 == nr_hugepages + surp + resv )) 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@109 -- # (( 1024 == nr_hugepages )) 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@17 -- # local get=HugePages_Total 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@18 -- # local node= 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@19 -- # local var val 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175638092 kB' 'MemAvailable: 178509256 kB' 'Buffers: 3896 kB' 'Cached: 10112724 kB' 'SwapCached: 0 kB' 'Active: 7138560 kB' 'Inactive: 3507524 kB' 'Active(anon): 6746552 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 532724 kB' 'Mapped: 170052 kB' 'Shmem: 6217088 kB' 'KReclaimable: 232164 kB' 'Slab: 811104 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 578940 kB' 'KernelStack: 20624 kB' 'PageTables: 8960 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103030604 kB' 'Committed_AS: 8274744 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315500 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.211 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.212 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # echo 1024 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # return 0 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@110 -- # (( 1024 == nr_hugepages + surp + resv )) 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@112 -- # get_nodes 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@27 -- # local node 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=512 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=512 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@32 -- # no_nodes=2 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@18 -- # local node=0 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@19 -- # local var val 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 97662684 kB' 'MemFree: 86910856 kB' 'MemUsed: 10751828 kB' 'SwapCached: 0 kB' 'Active: 4907776 kB' 'Inactive: 3335920 kB' 'Active(anon): 4750236 kB' 'Inactive(anon): 0 kB' 'Active(file): 157540 kB' 'Inactive(file): 3335920 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8121552 kB' 'Mapped: 89428 kB' 'AnonPages: 125328 kB' 'Shmem: 4628092 kB' 'KernelStack: 11528 kB' 'PageTables: 4128 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 124964 kB' 'Slab: 393984 kB' 'SReclaimable: 124964 kB' 'SUnreclaim: 269020 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 512' 'HugePages_Free: 512' 'HugePages_Surp: 0' 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.213 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.214 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # echo 0 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # return 0 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 1 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@18 -- # local node=1 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@19 -- # local var val 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node1/meminfo ]] 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node1/meminfo 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 93718468 kB' 'MemFree: 88727488 kB' 'MemUsed: 4990980 kB' 'SwapCached: 0 kB' 'Active: 2230828 kB' 'Inactive: 171604 kB' 'Active(anon): 1996360 kB' 'Inactive(anon): 0 kB' 'Active(file): 234468 kB' 'Inactive(file): 171604 kB' 'Unevictable: 0 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 1995072 kB' 'Mapped: 80624 kB' 'AnonPages: 407428 kB' 'Shmem: 1589000 kB' 'KernelStack: 9112 kB' 'PageTables: 4884 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 107200 kB' 'Slab: 417120 kB' 'SReclaimable: 107200 kB' 'SUnreclaim: 309920 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 512' 'HugePages_Free: 512' 'HugePages_Surp: 0' 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.215 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # continue 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # echo 0 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/common.sh@33 -- # return 0 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@128 -- # echo 'node0=512 expecting 512' 00:03:06.216 node0=512 expecting 512 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@128 -- # echo 'node1=512 expecting 512' 00:03:06.216 node1=512 expecting 512 00:03:06.216 17:48:59 setup.sh.hugepages.even_2G_alloc -- setup/hugepages.sh@130 -- # [[ 512 == \5\1\2 ]] 00:03:06.216 00:03:06.216 real 0m2.917s 00:03:06.216 user 0m1.174s 00:03:06.217 sys 0m1.797s 00:03:06.217 17:48:59 setup.sh.hugepages.even_2G_alloc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:03:06.217 17:48:59 setup.sh.hugepages.even_2G_alloc -- common/autotest_common.sh@10 -- # set +x 00:03:06.217 ************************************ 00:03:06.217 END TEST even_2G_alloc 00:03:06.217 ************************************ 00:03:06.217 17:48:59 setup.sh.hugepages -- common/autotest_common.sh@1142 -- # return 0 00:03:06.217 17:48:59 setup.sh.hugepages -- setup/hugepages.sh@213 -- # run_test odd_alloc odd_alloc 00:03:06.217 17:48:59 setup.sh.hugepages -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:03:06.217 17:48:59 setup.sh.hugepages -- common/autotest_common.sh@1105 -- # xtrace_disable 00:03:06.217 17:48:59 setup.sh.hugepages -- common/autotest_common.sh@10 -- # set +x 00:03:06.217 ************************************ 00:03:06.217 START TEST odd_alloc 00:03:06.217 ************************************ 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- common/autotest_common.sh@1123 -- # odd_alloc 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@159 -- # get_test_nr_hugepages 2098176 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@49 -- # local size=2098176 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@50 -- # (( 1 > 1 )) 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@57 -- # nr_hugepages=1025 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@62 -- # user_nodes=() 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@62 -- # local user_nodes 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@64 -- # local _nr_hugepages=1025 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@67 -- # nodes_test=() 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@67 -- # local -g nodes_test 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@69 -- # (( 0 > 0 )) 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@74 -- # (( 0 > 0 )) 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@82 -- # nodes_test[_no_nodes - 1]=512 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@83 -- # : 513 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@84 -- # : 1 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@82 -- # nodes_test[_no_nodes - 1]=513 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@83 -- # : 0 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@84 -- # : 0 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@160 -- # HUGEMEM=2049 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@160 -- # HUGE_EVEN_ALLOC=yes 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@160 -- # setup output 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/common.sh@9 -- # [[ output == output ]] 00:03:06.217 17:48:59 setup.sh.hugepages.odd_alloc -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:03:08.752 0000:00:04.7 (8086 2021): Already using the vfio-pci driver 00:03:08.752 0000:5e:00.0 (8086 0a54): Already using the vfio-pci driver 00:03:08.752 0000:00:04.6 (8086 2021): Already using the vfio-pci driver 00:03:08.752 0000:00:04.5 (8086 2021): Already using the vfio-pci driver 00:03:08.752 0000:00:04.4 (8086 2021): Already using the vfio-pci driver 00:03:08.752 0000:00:04.3 (8086 2021): Already using the vfio-pci driver 00:03:08.752 0000:00:04.2 (8086 2021): Already using the vfio-pci driver 00:03:08.752 0000:00:04.1 (8086 2021): Already using the vfio-pci driver 00:03:08.752 0000:00:04.0 (8086 2021): Already using the vfio-pci driver 00:03:09.012 0000:80:04.7 (8086 2021): Already using the vfio-pci driver 00:03:09.012 0000:80:04.6 (8086 2021): Already using the vfio-pci driver 00:03:09.012 0000:80:04.5 (8086 2021): Already using the vfio-pci driver 00:03:09.012 0000:80:04.4 (8086 2021): Already using the vfio-pci driver 00:03:09.012 0000:80:04.3 (8086 2021): Already using the vfio-pci driver 00:03:09.012 0000:80:04.2 (8086 2021): Already using the vfio-pci driver 00:03:09.012 0000:80:04.1 (8086 2021): Already using the vfio-pci driver 00:03:09.012 0000:80:04.0 (8086 2021): Already using the vfio-pci driver 00:03:09.012 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@161 -- # verify_nr_hugepages 00:03:09.012 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@89 -- # local node 00:03:09.012 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@90 -- # local sorted_t 00:03:09.012 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@91 -- # local sorted_s 00:03:09.012 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@92 -- # local surp 00:03:09.012 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@93 -- # local resv 00:03:09.012 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@94 -- # local anon 00:03:09.012 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:03:09.012 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:03:09.012 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@17 -- # local get=AnonHugePages 00:03:09.012 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@18 -- # local node= 00:03:09.012 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@19 -- # local var val 00:03:09.012 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:09.012 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:09.012 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:09.012 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:09.012 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:09.012 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:09.012 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.012 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175612396 kB' 'MemAvailable: 178483560 kB' 'Buffers: 3896 kB' 'Cached: 10112824 kB' 'SwapCached: 0 kB' 'Active: 7140160 kB' 'Inactive: 3507524 kB' 'Active(anon): 6748152 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 533796 kB' 'Mapped: 170156 kB' 'Shmem: 6217188 kB' 'KReclaimable: 232164 kB' 'Slab: 810628 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 578464 kB' 'KernelStack: 20528 kB' 'PageTables: 8664 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103029580 kB' 'Committed_AS: 8275400 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315468 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1025' 'HugePages_Free: 1025' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2099200 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.013 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # echo 0 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # return 0 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@97 -- # anon=0 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@18 -- # local node= 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@19 -- # local var val 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175613184 kB' 'MemAvailable: 178484348 kB' 'Buffers: 3896 kB' 'Cached: 10112828 kB' 'SwapCached: 0 kB' 'Active: 7139364 kB' 'Inactive: 3507524 kB' 'Active(anon): 6747356 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 533496 kB' 'Mapped: 170080 kB' 'Shmem: 6217192 kB' 'KReclaimable: 232164 kB' 'Slab: 810600 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 578436 kB' 'KernelStack: 20496 kB' 'PageTables: 8560 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103029580 kB' 'Committed_AS: 8275420 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315452 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1025' 'HugePages_Free: 1025' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2099200 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.014 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.015 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # echo 0 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # return 0 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@99 -- # surp=0 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@18 -- # local node= 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@19 -- # local var val 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175613184 kB' 'MemAvailable: 178484348 kB' 'Buffers: 3896 kB' 'Cached: 10112852 kB' 'SwapCached: 0 kB' 'Active: 7139372 kB' 'Inactive: 3507524 kB' 'Active(anon): 6747364 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 533500 kB' 'Mapped: 170080 kB' 'Shmem: 6217216 kB' 'KReclaimable: 232164 kB' 'Slab: 810600 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 578436 kB' 'KernelStack: 20496 kB' 'PageTables: 8560 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103029580 kB' 'Committed_AS: 8275440 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315452 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1025' 'HugePages_Free: 1025' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2099200 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.016 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:09.017 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # echo 0 00:03:09.018 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # return 0 00:03:09.018 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@100 -- # resv=0 00:03:09.018 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@102 -- # echo nr_hugepages=1025 00:03:09.018 nr_hugepages=1025 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:03:09.282 resv_hugepages=0 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:03:09.282 surplus_hugepages=0 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:03:09.282 anon_hugepages=0 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@107 -- # (( 1025 == nr_hugepages + surp + resv )) 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@109 -- # (( 1025 == nr_hugepages )) 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@17 -- # local get=HugePages_Total 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@18 -- # local node= 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@19 -- # local var val 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175614268 kB' 'MemAvailable: 178485432 kB' 'Buffers: 3896 kB' 'Cached: 10112864 kB' 'SwapCached: 0 kB' 'Active: 7139400 kB' 'Inactive: 3507524 kB' 'Active(anon): 6747392 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 533496 kB' 'Mapped: 170080 kB' 'Shmem: 6217228 kB' 'KReclaimable: 232164 kB' 'Slab: 810600 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 578436 kB' 'KernelStack: 20496 kB' 'PageTables: 8560 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103029580 kB' 'Committed_AS: 8275460 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315452 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1025' 'HugePages_Free: 1025' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2099200 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.282 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.283 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # echo 1025 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # return 0 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@110 -- # (( 1025 == nr_hugepages + surp + resv )) 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@112 -- # get_nodes 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@27 -- # local node 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=512 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=513 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@32 -- # no_nodes=2 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@18 -- # local node=0 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@19 -- # local var val 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 97662684 kB' 'MemFree: 86890512 kB' 'MemUsed: 10772172 kB' 'SwapCached: 0 kB' 'Active: 4908720 kB' 'Inactive: 3335920 kB' 'Active(anon): 4751180 kB' 'Inactive(anon): 0 kB' 'Active(file): 157540 kB' 'Inactive(file): 3335920 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8121572 kB' 'Mapped: 89456 kB' 'AnonPages: 126332 kB' 'Shmem: 4628112 kB' 'KernelStack: 11432 kB' 'PageTables: 3824 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 124964 kB' 'Slab: 393520 kB' 'SReclaimable: 124964 kB' 'SUnreclaim: 268556 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 512' 'HugePages_Free: 512' 'HugePages_Surp: 0' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.284 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # echo 0 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # return 0 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 1 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@18 -- # local node=1 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@19 -- # local var val 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node1/meminfo ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node1/meminfo 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 93718468 kB' 'MemFree: 88723252 kB' 'MemUsed: 4995216 kB' 'SwapCached: 0 kB' 'Active: 2230696 kB' 'Inactive: 171604 kB' 'Active(anon): 1996228 kB' 'Inactive(anon): 0 kB' 'Active(file): 234468 kB' 'Inactive(file): 171604 kB' 'Unevictable: 0 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 1995232 kB' 'Mapped: 80624 kB' 'AnonPages: 407160 kB' 'Shmem: 1589160 kB' 'KernelStack: 9080 kB' 'PageTables: 4736 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 107200 kB' 'Slab: 417080 kB' 'SReclaimable: 107200 kB' 'SUnreclaim: 309880 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 513' 'HugePages_Free: 513' 'HugePages_Surp: 0' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.285 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.286 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # continue 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # echo 0 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/common.sh@33 -- # return 0 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@128 -- # echo 'node0=512 expecting 513' 00:03:09.287 node0=512 expecting 513 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@128 -- # echo 'node1=513 expecting 512' 00:03:09.287 node1=513 expecting 512 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- setup/hugepages.sh@130 -- # [[ 512 513 == \5\1\2\ \5\1\3 ]] 00:03:09.287 00:03:09.287 real 0m2.936s 00:03:09.287 user 0m1.205s 00:03:09.287 sys 0m1.781s 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:03:09.287 17:49:02 setup.sh.hugepages.odd_alloc -- common/autotest_common.sh@10 -- # set +x 00:03:09.287 ************************************ 00:03:09.287 END TEST odd_alloc 00:03:09.287 ************************************ 00:03:09.287 17:49:02 setup.sh.hugepages -- common/autotest_common.sh@1142 -- # return 0 00:03:09.287 17:49:02 setup.sh.hugepages -- setup/hugepages.sh@214 -- # run_test custom_alloc custom_alloc 00:03:09.287 17:49:02 setup.sh.hugepages -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:03:09.287 17:49:02 setup.sh.hugepages -- common/autotest_common.sh@1105 -- # xtrace_disable 00:03:09.287 17:49:02 setup.sh.hugepages -- common/autotest_common.sh@10 -- # set +x 00:03:09.287 ************************************ 00:03:09.287 START TEST custom_alloc 00:03:09.287 ************************************ 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- common/autotest_common.sh@1123 -- # custom_alloc 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@167 -- # local IFS=, 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@169 -- # local node 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@170 -- # nodes_hp=() 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@170 -- # local nodes_hp 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@172 -- # local nr_hugepages=0 _nr_hugepages=0 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@174 -- # get_test_nr_hugepages 1048576 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@49 -- # local size=1048576 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@50 -- # (( 1 > 1 )) 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@57 -- # nr_hugepages=512 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@62 -- # user_nodes=() 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@62 -- # local user_nodes 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@64 -- # local _nr_hugepages=512 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@67 -- # nodes_test=() 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@67 -- # local -g nodes_test 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@69 -- # (( 0 > 0 )) 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@74 -- # (( 0 > 0 )) 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@82 -- # nodes_test[_no_nodes - 1]=256 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@83 -- # : 256 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@84 -- # : 1 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@82 -- # nodes_test[_no_nodes - 1]=256 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@83 -- # : 0 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@84 -- # : 0 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@81 -- # (( _no_nodes > 0 )) 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@175 -- # nodes_hp[0]=512 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@176 -- # (( 2 > 1 )) 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@177 -- # get_test_nr_hugepages 2097152 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@49 -- # local size=2097152 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@50 -- # (( 1 > 1 )) 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@57 -- # nr_hugepages=1024 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@62 -- # user_nodes=() 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@62 -- # local user_nodes 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@64 -- # local _nr_hugepages=1024 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@67 -- # nodes_test=() 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@67 -- # local -g nodes_test 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@69 -- # (( 0 > 0 )) 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@74 -- # (( 1 > 0 )) 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@75 -- # for _no_nodes in "${!nodes_hp[@]}" 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@76 -- # nodes_test[_no_nodes]=512 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@78 -- # return 0 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@178 -- # nodes_hp[1]=1024 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@181 -- # for node in "${!nodes_hp[@]}" 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@182 -- # HUGENODE+=("nodes_hp[$node]=${nodes_hp[node]}") 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@183 -- # (( _nr_hugepages += nodes_hp[node] )) 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@181 -- # for node in "${!nodes_hp[@]}" 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@182 -- # HUGENODE+=("nodes_hp[$node]=${nodes_hp[node]}") 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@183 -- # (( _nr_hugepages += nodes_hp[node] )) 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@186 -- # get_test_nr_hugepages_per_node 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@62 -- # user_nodes=() 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@62 -- # local user_nodes 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@64 -- # local _nr_hugepages=1024 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@67 -- # nodes_test=() 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@67 -- # local -g nodes_test 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@69 -- # (( 0 > 0 )) 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@74 -- # (( 2 > 0 )) 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@75 -- # for _no_nodes in "${!nodes_hp[@]}" 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@76 -- # nodes_test[_no_nodes]=512 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@75 -- # for _no_nodes in "${!nodes_hp[@]}" 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@76 -- # nodes_test[_no_nodes]=1024 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@78 -- # return 0 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@187 -- # HUGENODE='nodes_hp[0]=512,nodes_hp[1]=1024' 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@187 -- # setup output 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/common.sh@9 -- # [[ output == output ]] 00:03:09.287 17:49:02 setup.sh.hugepages.custom_alloc -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:03:11.879 0000:00:04.7 (8086 2021): Already using the vfio-pci driver 00:03:11.879 0000:5e:00.0 (8086 0a54): Already using the vfio-pci driver 00:03:11.879 0000:00:04.6 (8086 2021): Already using the vfio-pci driver 00:03:11.880 0000:00:04.5 (8086 2021): Already using the vfio-pci driver 00:03:11.880 0000:00:04.4 (8086 2021): Already using the vfio-pci driver 00:03:11.880 0000:00:04.3 (8086 2021): Already using the vfio-pci driver 00:03:11.880 0000:00:04.2 (8086 2021): Already using the vfio-pci driver 00:03:11.880 0000:00:04.1 (8086 2021): Already using the vfio-pci driver 00:03:11.880 0000:00:04.0 (8086 2021): Already using the vfio-pci driver 00:03:11.880 0000:80:04.7 (8086 2021): Already using the vfio-pci driver 00:03:11.880 0000:80:04.6 (8086 2021): Already using the vfio-pci driver 00:03:11.880 0000:80:04.5 (8086 2021): Already using the vfio-pci driver 00:03:11.880 0000:80:04.4 (8086 2021): Already using the vfio-pci driver 00:03:11.880 0000:80:04.3 (8086 2021): Already using the vfio-pci driver 00:03:11.880 0000:80:04.2 (8086 2021): Already using the vfio-pci driver 00:03:11.880 0000:80:04.1 (8086 2021): Already using the vfio-pci driver 00:03:11.880 0000:80:04.0 (8086 2021): Already using the vfio-pci driver 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@188 -- # nr_hugepages=1536 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@188 -- # verify_nr_hugepages 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@89 -- # local node 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@90 -- # local sorted_t 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@91 -- # local sorted_s 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@92 -- # local surp 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@93 -- # local resv 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@94 -- # local anon 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@17 -- # local get=AnonHugePages 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@18 -- # local node= 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@19 -- # local var val 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.144 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 174546272 kB' 'MemAvailable: 177417436 kB' 'Buffers: 3896 kB' 'Cached: 10112980 kB' 'SwapCached: 0 kB' 'Active: 7141572 kB' 'Inactive: 3507524 kB' 'Active(anon): 6749564 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 535428 kB' 'Mapped: 170500 kB' 'Shmem: 6217344 kB' 'KReclaimable: 232164 kB' 'Slab: 811312 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 579148 kB' 'KernelStack: 20688 kB' 'PageTables: 9288 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 102506316 kB' 'Committed_AS: 8278684 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315660 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1536' 'HugePages_Free: 1536' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 3145728 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.145 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # echo 0 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # return 0 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@97 -- # anon=0 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@18 -- # local node= 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@19 -- # local var val 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 174547436 kB' 'MemAvailable: 177418600 kB' 'Buffers: 3896 kB' 'Cached: 10112984 kB' 'SwapCached: 0 kB' 'Active: 7141760 kB' 'Inactive: 3507524 kB' 'Active(anon): 6749752 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 535688 kB' 'Mapped: 170196 kB' 'Shmem: 6217348 kB' 'KReclaimable: 232164 kB' 'Slab: 811652 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 579488 kB' 'KernelStack: 20896 kB' 'PageTables: 9724 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 102506316 kB' 'Committed_AS: 8278704 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315660 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1536' 'HugePages_Free: 1536' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 3145728 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.146 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.147 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # echo 0 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # return 0 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@99 -- # surp=0 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@18 -- # local node= 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@19 -- # local var val 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 174549564 kB' 'MemAvailable: 177420728 kB' 'Buffers: 3896 kB' 'Cached: 10113000 kB' 'SwapCached: 0 kB' 'Active: 7140376 kB' 'Inactive: 3507524 kB' 'Active(anon): 6748368 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 534248 kB' 'Mapped: 170136 kB' 'Shmem: 6217364 kB' 'KReclaimable: 232164 kB' 'Slab: 811696 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 579532 kB' 'KernelStack: 20544 kB' 'PageTables: 8656 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 102506316 kB' 'Committed_AS: 8278724 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315660 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1536' 'HugePages_Free: 1536' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 3145728 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.148 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.149 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # echo 0 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # return 0 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@100 -- # resv=0 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@102 -- # echo nr_hugepages=1536 00:03:12.150 nr_hugepages=1536 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:03:12.150 resv_hugepages=0 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:03:12.150 surplus_hugepages=0 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:03:12.150 anon_hugepages=0 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@107 -- # (( 1536 == nr_hugepages + surp + resv )) 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@109 -- # (( 1536 == nr_hugepages )) 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@17 -- # local get=HugePages_Total 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@18 -- # local node= 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@19 -- # local var val 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 174550468 kB' 'MemAvailable: 177421632 kB' 'Buffers: 3896 kB' 'Cached: 10113024 kB' 'SwapCached: 0 kB' 'Active: 7140448 kB' 'Inactive: 3507524 kB' 'Active(anon): 6748440 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 533784 kB' 'Mapped: 170092 kB' 'Shmem: 6217388 kB' 'KReclaimable: 232164 kB' 'Slab: 811720 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 579556 kB' 'KernelStack: 20528 kB' 'PageTables: 8556 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 102506316 kB' 'Committed_AS: 8277252 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315596 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1536' 'HugePages_Free: 1536' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 3145728 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.150 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.151 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # echo 1536 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # return 0 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@110 -- # (( 1536 == nr_hugepages + surp + resv )) 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@112 -- # get_nodes 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@27 -- # local node 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=512 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=1024 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@32 -- # no_nodes=2 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@18 -- # local node=0 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@19 -- # local var val 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 97662684 kB' 'MemFree: 86872768 kB' 'MemUsed: 10789916 kB' 'SwapCached: 0 kB' 'Active: 4910748 kB' 'Inactive: 3335920 kB' 'Active(anon): 4753208 kB' 'Inactive(anon): 0 kB' 'Active(file): 157540 kB' 'Inactive(file): 3335920 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8121608 kB' 'Mapped: 89468 kB' 'AnonPages: 128292 kB' 'Shmem: 4628148 kB' 'KernelStack: 11800 kB' 'PageTables: 4876 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 124964 kB' 'Slab: 394248 kB' 'SReclaimable: 124964 kB' 'SUnreclaim: 269284 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 512' 'HugePages_Free: 512' 'HugePages_Surp: 0' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.152 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.153 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # echo 0 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # return 0 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 1 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@18 -- # local node=1 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@19 -- # local var val 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node1/meminfo ]] 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node1/meminfo 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:12.154 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:12.413 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.413 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.413 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 93718468 kB' 'MemFree: 87675524 kB' 'MemUsed: 6042944 kB' 'SwapCached: 0 kB' 'Active: 2230564 kB' 'Inactive: 171604 kB' 'Active(anon): 1996096 kB' 'Inactive(anon): 0 kB' 'Active(file): 234468 kB' 'Inactive(file): 171604 kB' 'Unevictable: 0 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 1995332 kB' 'Mapped: 80624 kB' 'AnonPages: 406872 kB' 'Shmem: 1589260 kB' 'KernelStack: 9000 kB' 'PageTables: 4520 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 107200 kB' 'Slab: 417464 kB' 'SReclaimable: 107200 kB' 'SUnreclaim: 310264 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Surp: 0' 00:03:12.413 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.413 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.413 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.413 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.413 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.413 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.413 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.413 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.413 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.413 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.413 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.413 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.413 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.413 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.413 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.413 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.413 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # continue 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # echo 0 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/common.sh@33 -- # return 0 00:03:12.414 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:03:12.415 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:03:12.415 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:03:12.415 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:03:12.415 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@128 -- # echo 'node0=512 expecting 512' 00:03:12.415 node0=512 expecting 512 00:03:12.415 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:03:12.415 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:03:12.415 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:03:12.415 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@128 -- # echo 'node1=1024 expecting 1024' 00:03:12.415 node1=1024 expecting 1024 00:03:12.415 17:49:05 setup.sh.hugepages.custom_alloc -- setup/hugepages.sh@130 -- # [[ 512,1024 == \5\1\2\,\1\0\2\4 ]] 00:03:12.415 00:03:12.415 real 0m3.006s 00:03:12.415 user 0m1.287s 00:03:12.415 sys 0m1.783s 00:03:12.415 17:49:05 setup.sh.hugepages.custom_alloc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:03:12.415 17:49:05 setup.sh.hugepages.custom_alloc -- common/autotest_common.sh@10 -- # set +x 00:03:12.415 ************************************ 00:03:12.415 END TEST custom_alloc 00:03:12.415 ************************************ 00:03:12.415 17:49:05 setup.sh.hugepages -- common/autotest_common.sh@1142 -- # return 0 00:03:12.415 17:49:05 setup.sh.hugepages -- setup/hugepages.sh@215 -- # run_test no_shrink_alloc no_shrink_alloc 00:03:12.415 17:49:05 setup.sh.hugepages -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:03:12.415 17:49:05 setup.sh.hugepages -- common/autotest_common.sh@1105 -- # xtrace_disable 00:03:12.415 17:49:05 setup.sh.hugepages -- common/autotest_common.sh@10 -- # set +x 00:03:12.415 ************************************ 00:03:12.415 START TEST no_shrink_alloc 00:03:12.415 ************************************ 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- common/autotest_common.sh@1123 -- # no_shrink_alloc 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@195 -- # get_test_nr_hugepages 2097152 0 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@49 -- # local size=2097152 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@50 -- # (( 2 > 1 )) 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@51 -- # shift 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@52 -- # node_ids=('0') 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@52 -- # local node_ids 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@55 -- # (( size >= default_hugepages )) 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@57 -- # nr_hugepages=1024 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@58 -- # get_test_nr_hugepages_per_node 0 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@62 -- # user_nodes=('0') 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@62 -- # local user_nodes 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@64 -- # local _nr_hugepages=1024 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@65 -- # local _no_nodes=2 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@67 -- # nodes_test=() 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@67 -- # local -g nodes_test 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@69 -- # (( 1 > 0 )) 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@70 -- # for _no_nodes in "${user_nodes[@]}" 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@71 -- # nodes_test[_no_nodes]=1024 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@73 -- # return 0 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@198 -- # setup output 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@9 -- # [[ output == output ]] 00:03:12.415 17:49:05 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:03:14.942 0000:00:04.7 (8086 2021): Already using the vfio-pci driver 00:03:14.942 0000:5e:00.0 (8086 0a54): Already using the vfio-pci driver 00:03:14.942 0000:00:04.6 (8086 2021): Already using the vfio-pci driver 00:03:14.942 0000:00:04.5 (8086 2021): Already using the vfio-pci driver 00:03:14.942 0000:00:04.4 (8086 2021): Already using the vfio-pci driver 00:03:14.942 0000:00:04.3 (8086 2021): Already using the vfio-pci driver 00:03:14.942 0000:00:04.2 (8086 2021): Already using the vfio-pci driver 00:03:14.942 0000:00:04.1 (8086 2021): Already using the vfio-pci driver 00:03:14.942 0000:00:04.0 (8086 2021): Already using the vfio-pci driver 00:03:14.942 0000:80:04.7 (8086 2021): Already using the vfio-pci driver 00:03:14.942 0000:80:04.6 (8086 2021): Already using the vfio-pci driver 00:03:14.942 0000:80:04.5 (8086 2021): Already using the vfio-pci driver 00:03:14.942 0000:80:04.4 (8086 2021): Already using the vfio-pci driver 00:03:14.942 0000:80:04.3 (8086 2021): Already using the vfio-pci driver 00:03:14.942 0000:80:04.2 (8086 2021): Already using the vfio-pci driver 00:03:14.942 0000:80:04.1 (8086 2021): Already using the vfio-pci driver 00:03:14.942 0000:80:04.0 (8086 2021): Already using the vfio-pci driver 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@199 -- # verify_nr_hugepages 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@89 -- # local node 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@90 -- # local sorted_t 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@91 -- # local sorted_s 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@92 -- # local surp 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@93 -- # local resv 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@94 -- # local anon 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@17 -- # local get=AnonHugePages 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@18 -- # local node= 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@19 -- # local var val 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175606476 kB' 'MemAvailable: 178477640 kB' 'Buffers: 3896 kB' 'Cached: 10113136 kB' 'SwapCached: 0 kB' 'Active: 7140984 kB' 'Inactive: 3507524 kB' 'Active(anon): 6748976 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 534196 kB' 'Mapped: 170180 kB' 'Shmem: 6217500 kB' 'KReclaimable: 232164 kB' 'Slab: 811760 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 579596 kB' 'KernelStack: 20496 kB' 'PageTables: 8556 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103030604 kB' 'Committed_AS: 8276608 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315532 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.206 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # echo 0 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # return 0 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@97 -- # anon=0 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@18 -- # local node= 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@19 -- # local var val 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.207 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175609800 kB' 'MemAvailable: 178480964 kB' 'Buffers: 3896 kB' 'Cached: 10113148 kB' 'SwapCached: 0 kB' 'Active: 7140696 kB' 'Inactive: 3507524 kB' 'Active(anon): 6748688 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 534436 kB' 'Mapped: 170104 kB' 'Shmem: 6217512 kB' 'KReclaimable: 232164 kB' 'Slab: 811752 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 579588 kB' 'KernelStack: 20480 kB' 'PageTables: 8512 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103030604 kB' 'Committed_AS: 8276624 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315500 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.208 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # echo 0 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # return 0 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@99 -- # surp=0 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@18 -- # local node= 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@19 -- # local var val 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175610452 kB' 'MemAvailable: 178481616 kB' 'Buffers: 3896 kB' 'Cached: 10113156 kB' 'SwapCached: 0 kB' 'Active: 7140696 kB' 'Inactive: 3507524 kB' 'Active(anon): 6748688 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 534432 kB' 'Mapped: 170104 kB' 'Shmem: 6217520 kB' 'KReclaimable: 232164 kB' 'Slab: 811752 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 579588 kB' 'KernelStack: 20480 kB' 'PageTables: 8512 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103030604 kB' 'Committed_AS: 8276648 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315500 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.209 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.210 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # echo 0 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # return 0 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@100 -- # resv=0 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@102 -- # echo nr_hugepages=1024 00:03:15.211 nr_hugepages=1024 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:03:15.211 resv_hugepages=0 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:03:15.211 surplus_hugepages=0 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:03:15.211 anon_hugepages=0 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@107 -- # (( 1024 == nr_hugepages + surp + resv )) 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@109 -- # (( 1024 == nr_hugepages )) 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@17 -- # local get=HugePages_Total 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@18 -- # local node= 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@19 -- # local var val 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175616160 kB' 'MemAvailable: 178487324 kB' 'Buffers: 3896 kB' 'Cached: 10113176 kB' 'SwapCached: 0 kB' 'Active: 7140576 kB' 'Inactive: 3507524 kB' 'Active(anon): 6748568 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 534264 kB' 'Mapped: 170104 kB' 'Shmem: 6217540 kB' 'KReclaimable: 232164 kB' 'Slab: 811736 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 579572 kB' 'KernelStack: 20464 kB' 'PageTables: 8460 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103030604 kB' 'Committed_AS: 8276672 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315500 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.211 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.212 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # echo 1024 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # return 0 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@110 -- # (( 1024 == nr_hugepages + surp + resv )) 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@112 -- # get_nodes 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@27 -- # local node 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=1024 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=0 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@32 -- # no_nodes=2 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@18 -- # local node=0 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@19 -- # local var val 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 97662684 kB' 'MemFree: 85824128 kB' 'MemUsed: 11838556 kB' 'SwapCached: 0 kB' 'Active: 4909476 kB' 'Inactive: 3335920 kB' 'Active(anon): 4751936 kB' 'Inactive(anon): 0 kB' 'Active(file): 157540 kB' 'Inactive(file): 3335920 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8121632 kB' 'Mapped: 89480 kB' 'AnonPages: 126860 kB' 'Shmem: 4628172 kB' 'KernelStack: 11384 kB' 'PageTables: 3688 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 124964 kB' 'Slab: 393932 kB' 'SReclaimable: 124964 kB' 'SUnreclaim: 268968 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Surp: 0' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.213 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # echo 0 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # return 0 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:03:15.214 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@128 -- # echo 'node0=1024 expecting 1024' 00:03:15.216 node0=1024 expecting 1024 00:03:15.216 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@130 -- # [[ 1024 == \1\0\2\4 ]] 00:03:15.216 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@202 -- # CLEAR_HUGE=no 00:03:15.216 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@202 -- # NRHUGE=512 00:03:15.216 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@202 -- # setup output 00:03:15.216 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@9 -- # [[ output == output ]] 00:03:15.216 17:49:08 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:03:17.752 0000:00:04.7 (8086 2021): Already using the vfio-pci driver 00:03:17.752 0000:5e:00.0 (8086 0a54): Already using the vfio-pci driver 00:03:17.752 0000:00:04.6 (8086 2021): Already using the vfio-pci driver 00:03:17.752 0000:00:04.5 (8086 2021): Already using the vfio-pci driver 00:03:17.752 0000:00:04.4 (8086 2021): Already using the vfio-pci driver 00:03:17.752 0000:00:04.3 (8086 2021): Already using the vfio-pci driver 00:03:17.752 0000:00:04.2 (8086 2021): Already using the vfio-pci driver 00:03:17.752 0000:00:04.1 (8086 2021): Already using the vfio-pci driver 00:03:17.752 0000:00:04.0 (8086 2021): Already using the vfio-pci driver 00:03:17.752 0000:80:04.7 (8086 2021): Already using the vfio-pci driver 00:03:17.752 0000:80:04.6 (8086 2021): Already using the vfio-pci driver 00:03:17.752 0000:80:04.5 (8086 2021): Already using the vfio-pci driver 00:03:17.752 0000:80:04.4 (8086 2021): Already using the vfio-pci driver 00:03:17.752 0000:80:04.3 (8086 2021): Already using the vfio-pci driver 00:03:17.752 0000:80:04.2 (8086 2021): Already using the vfio-pci driver 00:03:17.752 0000:80:04.1 (8086 2021): Already using the vfio-pci driver 00:03:17.752 0000:80:04.0 (8086 2021): Already using the vfio-pci driver 00:03:17.752 INFO: Requested 512 hugepages but 1024 already allocated on node0 00:03:17.752 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@204 -- # verify_nr_hugepages 00:03:17.752 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@89 -- # local node 00:03:17.752 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@90 -- # local sorted_t 00:03:17.752 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@91 -- # local sorted_s 00:03:17.752 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@92 -- # local surp 00:03:17.752 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@93 -- # local resv 00:03:17.752 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@94 -- # local anon 00:03:17.752 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@96 -- # [[ always [madvise] never != *\[\n\e\v\e\r\]* ]] 00:03:17.752 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@97 -- # get_meminfo AnonHugePages 00:03:17.752 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@17 -- # local get=AnonHugePages 00:03:17.752 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@18 -- # local node= 00:03:17.752 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@19 -- # local var val 00:03:17.752 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:17.752 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:17.752 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:17.752 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:17.752 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:17.752 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:17.752 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.752 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175595792 kB' 'MemAvailable: 178466956 kB' 'Buffers: 3896 kB' 'Cached: 10113264 kB' 'SwapCached: 0 kB' 'Active: 7143724 kB' 'Inactive: 3507524 kB' 'Active(anon): 6751716 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 536880 kB' 'Mapped: 170136 kB' 'Shmem: 6217628 kB' 'KReclaimable: 232164 kB' 'Slab: 811276 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 579112 kB' 'KernelStack: 20848 kB' 'PageTables: 9716 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103030604 kB' 'Committed_AS: 8279988 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315708 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Buffers == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Cached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapCached == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unevictable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mlocked == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapFree == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswap == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Dirty == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Writeback == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonPages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mapped == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Shmem == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Slab == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.753 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KernelStack == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ PageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Bounce == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Percpu == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \A\n\o\n\H\u\g\e\P\a\g\e\s ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # echo 0 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # return 0 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@97 -- # anon=0 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@99 -- # get_meminfo HugePages_Surp 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@18 -- # local node= 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@19 -- # local var val 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175595396 kB' 'MemAvailable: 178466560 kB' 'Buffers: 3896 kB' 'Cached: 10113264 kB' 'SwapCached: 0 kB' 'Active: 7142024 kB' 'Inactive: 3507524 kB' 'Active(anon): 6750016 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 535600 kB' 'Mapped: 170032 kB' 'Shmem: 6217628 kB' 'KReclaimable: 232164 kB' 'Slab: 811376 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 579212 kB' 'KernelStack: 20592 kB' 'PageTables: 9048 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103030604 kB' 'Committed_AS: 8279976 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315612 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.754 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.755 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # echo 0 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # return 0 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@99 -- # surp=0 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@100 -- # get_meminfo HugePages_Rsvd 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@17 -- # local get=HugePages_Rsvd 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@18 -- # local node= 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@19 -- # local var val 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175595808 kB' 'MemAvailable: 178466972 kB' 'Buffers: 3896 kB' 'Cached: 10113284 kB' 'SwapCached: 0 kB' 'Active: 7141388 kB' 'Inactive: 3507524 kB' 'Active(anon): 6749380 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 535004 kB' 'Mapped: 170108 kB' 'Shmem: 6217648 kB' 'KReclaimable: 232164 kB' 'Slab: 811600 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 579436 kB' 'KernelStack: 20496 kB' 'PageTables: 8532 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103030604 kB' 'Committed_AS: 8277384 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315532 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.756 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.757 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Rsvd == \H\u\g\e\P\a\g\e\s\_\R\s\v\d ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # echo 0 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # return 0 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@100 -- # resv=0 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@102 -- # echo nr_hugepages=1024 00:03:17.758 nr_hugepages=1024 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@103 -- # echo resv_hugepages=0 00:03:17.758 resv_hugepages=0 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@104 -- # echo surplus_hugepages=0 00:03:17.758 surplus_hugepages=0 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@105 -- # echo anon_hugepages=0 00:03:17.758 anon_hugepages=0 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@107 -- # (( 1024 == nr_hugepages + surp + resv )) 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@109 -- # (( 1024 == nr_hugepages )) 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@110 -- # get_meminfo HugePages_Total 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@17 -- # local get=HugePages_Total 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@18 -- # local node= 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@19 -- # local var val 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node/meminfo ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@25 -- # [[ -n '' ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 191381152 kB' 'MemFree: 175596364 kB' 'MemAvailable: 178467528 kB' 'Buffers: 3896 kB' 'Cached: 10113324 kB' 'SwapCached: 0 kB' 'Active: 7141088 kB' 'Inactive: 3507524 kB' 'Active(anon): 6749080 kB' 'Inactive(anon): 0 kB' 'Active(file): 392008 kB' 'Inactive(file): 3507524 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'SwapTotal: 8388604 kB' 'SwapFree: 8388604 kB' 'Zswap: 0 kB' 'Zswapped: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'AnonPages: 534656 kB' 'Mapped: 170108 kB' 'Shmem: 6217688 kB' 'KReclaimable: 232164 kB' 'Slab: 811600 kB' 'SReclaimable: 232164 kB' 'SUnreclaim: 579436 kB' 'KernelStack: 20480 kB' 'PageTables: 8480 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'CommitLimit: 103030604 kB' 'Committed_AS: 8277404 kB' 'VmallocTotal: 34359738367 kB' 'VmallocUsed: 315532 kB' 'VmallocChunk: 0 kB' 'Percpu: 77952 kB' 'HardwareCorrupted: 0 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'CmaTotal: 0 kB' 'CmaFree: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Rsvd: 0' 'HugePages_Surp: 0' 'Hugepagesize: 2048 kB' 'Hugetlb: 2097152 kB' 'DirectMap4k: 2978772 kB' 'DirectMap2M: 15575040 kB' 'DirectMap1G: 183500800 kB' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemAvailable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.758 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Buffers == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Cached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswap == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Zswapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CommitLimit == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Committed_AS == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.759 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocUsed == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ VmallocChunk == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Percpu == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HardwareCorrupted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaTotal == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ CmaFree == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\T\o\t\a\l ]] 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # echo 1024 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # return 0 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@110 -- # (( 1024 == nr_hugepages + surp + resv )) 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@112 -- # get_nodes 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@27 -- # local node 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=1024 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@29 -- # for node in /sys/devices/system/node/node+([0-9]) 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@30 -- # nodes_sys[${node##*node}]=0 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@32 -- # no_nodes=2 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@33 -- # (( no_nodes > 0 )) 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@115 -- # for node in "${!nodes_test[@]}" 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@116 -- # (( nodes_test[node] += resv )) 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@117 -- # get_meminfo HugePages_Surp 0 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@17 -- # local get=HugePages_Surp 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@18 -- # local node=0 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@19 -- # local var val 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@20 -- # local mem_f mem 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@22 -- # mem_f=/proc/meminfo 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@23 -- # [[ -e /sys/devices/system/node/node0/meminfo ]] 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@24 -- # mem_f=/sys/devices/system/node/node0/meminfo 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@28 -- # mapfile -t mem 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@29 -- # mem=("${mem[@]#Node +([0-9]) }") 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@16 -- # printf '%s\n' 'MemTotal: 97662684 kB' 'MemFree: 85815092 kB' 'MemUsed: 11847592 kB' 'SwapCached: 0 kB' 'Active: 4908224 kB' 'Inactive: 3335920 kB' 'Active(anon): 4750684 kB' 'Inactive(anon): 0 kB' 'Active(file): 157540 kB' 'Inactive(file): 3335920 kB' 'Unevictable: 3072 kB' 'Mlocked: 0 kB' 'Dirty: 0 kB' 'Writeback: 0 kB' 'FilePages: 8121632 kB' 'Mapped: 89484 kB' 'AnonPages: 125672 kB' 'Shmem: 4628172 kB' 'KernelStack: 11384 kB' 'PageTables: 3696 kB' 'SecPageTables: 0 kB' 'NFS_Unstable: 0 kB' 'Bounce: 0 kB' 'WritebackTmp: 0 kB' 'KReclaimable: 124964 kB' 'Slab: 393824 kB' 'SReclaimable: 124964 kB' 'SUnreclaim: 268860 kB' 'AnonHugePages: 0 kB' 'ShmemHugePages: 0 kB' 'ShmemPmdMapped: 0 kB' 'FileHugePages: 0 kB' 'FilePmdMapped: 0 kB' 'Unaccepted: 0 kB' 'HugePages_Total: 1024' 'HugePages_Free: 1024' 'HugePages_Surp: 0' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemTotal == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemFree == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ MemUsed == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SwapCached == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:17.760 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(anon) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Active(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Inactive(file) == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unevictable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mlocked == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Dirty == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Writeback == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FilePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Mapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonPages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Shmem == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KernelStack == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.021 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ PageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SecPageTables == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ NFS_Unstable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Bounce == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ WritebackTmp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ KReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Slab == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SReclaimable == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ SUnreclaim == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ AnonHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ ShmemPmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FileHugePages == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ FilePmdMapped == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ Unaccepted == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Total == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Free == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # continue 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # IFS=': ' 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@31 -- # read -r var val _ 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@32 -- # [[ HugePages_Surp == \H\u\g\e\P\a\g\e\s\_\S\u\r\p ]] 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # echo 0 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/common.sh@33 -- # return 0 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@117 -- # (( nodes_test[node] += 0 )) 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@126 -- # for node in "${!nodes_test[@]}" 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@127 -- # sorted_t[nodes_test[node]]=1 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@127 -- # sorted_s[nodes_sys[node]]=1 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@128 -- # echo 'node0=1024 expecting 1024' 00:03:18.022 node0=1024 expecting 1024 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- setup/hugepages.sh@130 -- # [[ 1024 == \1\0\2\4 ]] 00:03:18.022 00:03:18.022 real 0m5.535s 00:03:18.022 user 0m2.191s 00:03:18.022 sys 0m3.385s 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:03:18.022 17:49:11 setup.sh.hugepages.no_shrink_alloc -- common/autotest_common.sh@10 -- # set +x 00:03:18.022 ************************************ 00:03:18.022 END TEST no_shrink_alloc 00:03:18.022 ************************************ 00:03:18.022 17:49:11 setup.sh.hugepages -- common/autotest_common.sh@1142 -- # return 0 00:03:18.022 17:49:11 setup.sh.hugepages -- setup/hugepages.sh@217 -- # clear_hp 00:03:18.022 17:49:11 setup.sh.hugepages -- setup/hugepages.sh@37 -- # local node hp 00:03:18.022 17:49:11 setup.sh.hugepages -- setup/hugepages.sh@39 -- # for node in "${!nodes_sys[@]}" 00:03:18.022 17:49:11 setup.sh.hugepages -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:03:18.022 17:49:11 setup.sh.hugepages -- setup/hugepages.sh@41 -- # echo 0 00:03:18.022 17:49:11 setup.sh.hugepages -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:03:18.022 17:49:11 setup.sh.hugepages -- setup/hugepages.sh@41 -- # echo 0 00:03:18.022 17:49:11 setup.sh.hugepages -- setup/hugepages.sh@39 -- # for node in "${!nodes_sys[@]}" 00:03:18.022 17:49:11 setup.sh.hugepages -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:03:18.022 17:49:11 setup.sh.hugepages -- setup/hugepages.sh@41 -- # echo 0 00:03:18.022 17:49:11 setup.sh.hugepages -- setup/hugepages.sh@40 -- # for hp in "/sys/devices/system/node/node$node/hugepages/hugepages-"* 00:03:18.022 17:49:11 setup.sh.hugepages -- setup/hugepages.sh@41 -- # echo 0 00:03:18.022 17:49:11 setup.sh.hugepages -- setup/hugepages.sh@45 -- # export CLEAR_HUGE=yes 00:03:18.022 17:49:11 setup.sh.hugepages -- setup/hugepages.sh@45 -- # CLEAR_HUGE=yes 00:03:18.022 00:03:18.022 real 0m21.827s 00:03:18.022 user 0m8.588s 00:03:18.022 sys 0m12.794s 00:03:18.022 17:49:11 setup.sh.hugepages -- common/autotest_common.sh@1124 -- # xtrace_disable 00:03:18.022 17:49:11 setup.sh.hugepages -- common/autotest_common.sh@10 -- # set +x 00:03:18.022 ************************************ 00:03:18.022 END TEST hugepages 00:03:18.022 ************************************ 00:03:18.022 17:49:11 setup.sh -- common/autotest_common.sh@1142 -- # return 0 00:03:18.022 17:49:11 setup.sh -- setup/test-setup.sh@14 -- # run_test driver /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/driver.sh 00:03:18.022 17:49:11 setup.sh -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:03:18.022 17:49:11 setup.sh -- common/autotest_common.sh@1105 -- # xtrace_disable 00:03:18.022 17:49:11 setup.sh -- common/autotest_common.sh@10 -- # set +x 00:03:18.022 ************************************ 00:03:18.022 START TEST driver 00:03:18.022 ************************************ 00:03:18.022 17:49:11 setup.sh.driver -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/driver.sh 00:03:18.022 * Looking for test storage... 00:03:18.022 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup 00:03:18.022 17:49:11 setup.sh.driver -- setup/driver.sh@68 -- # setup reset 00:03:18.022 17:49:11 setup.sh.driver -- setup/common.sh@9 -- # [[ reset == output ]] 00:03:18.022 17:49:11 setup.sh.driver -- setup/common.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:03:22.219 17:49:15 setup.sh.driver -- setup/driver.sh@69 -- # run_test guess_driver guess_driver 00:03:22.219 17:49:15 setup.sh.driver -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:03:22.219 17:49:15 setup.sh.driver -- common/autotest_common.sh@1105 -- # xtrace_disable 00:03:22.219 17:49:15 setup.sh.driver -- common/autotest_common.sh@10 -- # set +x 00:03:22.219 ************************************ 00:03:22.219 START TEST guess_driver 00:03:22.219 ************************************ 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- common/autotest_common.sh@1123 -- # guess_driver 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@46 -- # local driver setup_driver marker 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@47 -- # local fail=0 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@49 -- # pick_driver 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@36 -- # vfio 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@21 -- # local iommu_grups 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@22 -- # local unsafe_vfio 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@24 -- # [[ -e /sys/module/vfio/parameters/enable_unsafe_noiommu_mode ]] 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@25 -- # unsafe_vfio=N 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@27 -- # iommu_groups=(/sys/kernel/iommu_groups/*) 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@29 -- # (( 174 > 0 )) 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@30 -- # is_driver vfio_pci 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@14 -- # mod vfio_pci 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@12 -- # dep vfio_pci 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@11 -- # modprobe --show-depends vfio_pci 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@12 -- # [[ insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/virt/lib/irqbypass.ko.xz 00:03:22.219 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/iommu/iommufd/iommufd.ko.xz 00:03:22.219 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/vfio/vfio.ko.xz 00:03:22.219 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/iommu/iommufd/iommufd.ko.xz 00:03:22.219 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/vfio/vfio.ko.xz 00:03:22.219 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/vfio/vfio_iommu_type1.ko.xz 00:03:22.219 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/vfio/pci/vfio-pci-core.ko.xz 00:03:22.219 insmod /lib/modules/6.7.0-68.fc38.x86_64/kernel/drivers/vfio/pci/vfio-pci.ko.xz == *\.\k\o* ]] 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@30 -- # return 0 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@37 -- # echo vfio-pci 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@49 -- # driver=vfio-pci 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@51 -- # [[ vfio-pci == \N\o\ \v\a\l\i\d\ \d\r\i\v\e\r\ \f\o\u\n\d ]] 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@56 -- # echo 'Looking for driver=vfio-pci' 00:03:22.219 Looking for driver=vfio-pci 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/driver.sh@45 -- # setup output config 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/common.sh@9 -- # [[ output == output ]] 00:03:22.219 17:49:15 setup.sh.driver.guess_driver -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh config 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:03:24.754 17:49:18 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:03:25.322 17:49:19 setup.sh.driver.guess_driver -- setup/driver.sh@58 -- # [[ -> == \-\> ]] 00:03:25.322 17:49:19 setup.sh.driver.guess_driver -- setup/driver.sh@61 -- # [[ vfio-pci == vfio-pci ]] 00:03:25.322 17:49:19 setup.sh.driver.guess_driver -- setup/driver.sh@57 -- # read -r _ _ _ _ marker setup_driver 00:03:25.582 17:49:19 setup.sh.driver.guess_driver -- setup/driver.sh@64 -- # (( fail == 0 )) 00:03:25.582 17:49:19 setup.sh.driver.guess_driver -- setup/driver.sh@65 -- # setup reset 00:03:25.582 17:49:19 setup.sh.driver.guess_driver -- setup/common.sh@9 -- # [[ reset == output ]] 00:03:25.582 17:49:19 setup.sh.driver.guess_driver -- setup/common.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:03:29.774 00:03:29.774 real 0m7.210s 00:03:29.774 user 0m1.916s 00:03:29.774 sys 0m3.736s 00:03:29.774 17:49:22 setup.sh.driver.guess_driver -- common/autotest_common.sh@1124 -- # xtrace_disable 00:03:29.774 17:49:22 setup.sh.driver.guess_driver -- common/autotest_common.sh@10 -- # set +x 00:03:29.774 ************************************ 00:03:29.774 END TEST guess_driver 00:03:29.774 ************************************ 00:03:29.774 17:49:22 setup.sh.driver -- common/autotest_common.sh@1142 -- # return 0 00:03:29.774 00:03:29.774 real 0m11.220s 00:03:29.774 user 0m3.006s 00:03:29.774 sys 0m5.866s 00:03:29.774 17:49:22 setup.sh.driver -- common/autotest_common.sh@1124 -- # xtrace_disable 00:03:29.774 17:49:22 setup.sh.driver -- common/autotest_common.sh@10 -- # set +x 00:03:29.774 ************************************ 00:03:29.774 END TEST driver 00:03:29.774 ************************************ 00:03:29.774 17:49:22 setup.sh -- common/autotest_common.sh@1142 -- # return 0 00:03:29.774 17:49:22 setup.sh -- setup/test-setup.sh@15 -- # run_test devices /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/devices.sh 00:03:29.774 17:49:22 setup.sh -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:03:29.774 17:49:22 setup.sh -- common/autotest_common.sh@1105 -- # xtrace_disable 00:03:29.774 17:49:22 setup.sh -- common/autotest_common.sh@10 -- # set +x 00:03:29.774 ************************************ 00:03:29.774 START TEST devices 00:03:29.774 ************************************ 00:03:29.774 17:49:22 setup.sh.devices -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/devices.sh 00:03:29.774 * Looking for test storage... 00:03:29.774 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup 00:03:29.774 17:49:22 setup.sh.devices -- setup/devices.sh@190 -- # trap cleanup EXIT 00:03:29.774 17:49:22 setup.sh.devices -- setup/devices.sh@192 -- # setup reset 00:03:29.774 17:49:22 setup.sh.devices -- setup/common.sh@9 -- # [[ reset == output ]] 00:03:29.774 17:49:22 setup.sh.devices -- setup/common.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:03:32.341 17:49:25 setup.sh.devices -- setup/devices.sh@194 -- # get_zoned_devs 00:03:32.341 17:49:25 setup.sh.devices -- common/autotest_common.sh@1669 -- # zoned_devs=() 00:03:32.341 17:49:25 setup.sh.devices -- common/autotest_common.sh@1669 -- # local -gA zoned_devs 00:03:32.341 17:49:25 setup.sh.devices -- common/autotest_common.sh@1670 -- # local nvme bdf 00:03:32.341 17:49:25 setup.sh.devices -- common/autotest_common.sh@1672 -- # for nvme in /sys/block/nvme* 00:03:32.341 17:49:25 setup.sh.devices -- common/autotest_common.sh@1673 -- # is_block_zoned nvme0n1 00:03:32.341 17:49:25 setup.sh.devices -- common/autotest_common.sh@1662 -- # local device=nvme0n1 00:03:32.341 17:49:25 setup.sh.devices -- common/autotest_common.sh@1664 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:03:32.341 17:49:25 setup.sh.devices -- common/autotest_common.sh@1665 -- # [[ none != none ]] 00:03:32.341 17:49:25 setup.sh.devices -- setup/devices.sh@196 -- # blocks=() 00:03:32.341 17:49:25 setup.sh.devices -- setup/devices.sh@196 -- # declare -a blocks 00:03:32.341 17:49:25 setup.sh.devices -- setup/devices.sh@197 -- # blocks_to_pci=() 00:03:32.341 17:49:25 setup.sh.devices -- setup/devices.sh@197 -- # declare -A blocks_to_pci 00:03:32.341 17:49:25 setup.sh.devices -- setup/devices.sh@198 -- # min_disk_size=3221225472 00:03:32.341 17:49:25 setup.sh.devices -- setup/devices.sh@200 -- # for block in "/sys/block/nvme"!(*c*) 00:03:32.341 17:49:25 setup.sh.devices -- setup/devices.sh@201 -- # ctrl=nvme0n1 00:03:32.341 17:49:25 setup.sh.devices -- setup/devices.sh@201 -- # ctrl=nvme0 00:03:32.341 17:49:25 setup.sh.devices -- setup/devices.sh@202 -- # pci=0000:5e:00.0 00:03:32.341 17:49:25 setup.sh.devices -- setup/devices.sh@203 -- # [[ '' == *\0\0\0\0\:\5\e\:\0\0\.\0* ]] 00:03:32.341 17:49:25 setup.sh.devices -- setup/devices.sh@204 -- # block_in_use nvme0n1 00:03:32.341 17:49:25 setup.sh.devices -- scripts/common.sh@378 -- # local block=nvme0n1 pt 00:03:32.341 17:49:25 setup.sh.devices -- scripts/common.sh@387 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:03:32.341 No valid GPT data, bailing 00:03:32.341 17:49:25 setup.sh.devices -- scripts/common.sh@391 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:03:32.341 17:49:25 setup.sh.devices -- scripts/common.sh@391 -- # pt= 00:03:32.341 17:49:25 setup.sh.devices -- scripts/common.sh@392 -- # return 1 00:03:32.341 17:49:25 setup.sh.devices -- setup/devices.sh@204 -- # sec_size_to_bytes nvme0n1 00:03:32.341 17:49:25 setup.sh.devices -- setup/common.sh@76 -- # local dev=nvme0n1 00:03:32.341 17:49:25 setup.sh.devices -- setup/common.sh@78 -- # [[ -e /sys/block/nvme0n1 ]] 00:03:32.341 17:49:25 setup.sh.devices -- setup/common.sh@80 -- # echo 1000204886016 00:03:32.341 17:49:25 setup.sh.devices -- setup/devices.sh@204 -- # (( 1000204886016 >= min_disk_size )) 00:03:32.341 17:49:25 setup.sh.devices -- setup/devices.sh@205 -- # blocks+=("${block##*/}") 00:03:32.341 17:49:25 setup.sh.devices -- setup/devices.sh@206 -- # blocks_to_pci["${block##*/}"]=0000:5e:00.0 00:03:32.341 17:49:25 setup.sh.devices -- setup/devices.sh@209 -- # (( 1 > 0 )) 00:03:32.341 17:49:25 setup.sh.devices -- setup/devices.sh@211 -- # declare -r test_disk=nvme0n1 00:03:32.341 17:49:25 setup.sh.devices -- setup/devices.sh@213 -- # run_test nvme_mount nvme_mount 00:03:32.341 17:49:25 setup.sh.devices -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:03:32.341 17:49:25 setup.sh.devices -- common/autotest_common.sh@1105 -- # xtrace_disable 00:03:32.341 17:49:25 setup.sh.devices -- common/autotest_common.sh@10 -- # set +x 00:03:32.341 ************************************ 00:03:32.341 START TEST nvme_mount 00:03:32.341 ************************************ 00:03:32.341 17:49:25 setup.sh.devices.nvme_mount -- common/autotest_common.sh@1123 -- # nvme_mount 00:03:32.341 17:49:25 setup.sh.devices.nvme_mount -- setup/devices.sh@95 -- # nvme_disk=nvme0n1 00:03:32.341 17:49:25 setup.sh.devices.nvme_mount -- setup/devices.sh@96 -- # nvme_disk_p=nvme0n1p1 00:03:32.341 17:49:25 setup.sh.devices.nvme_mount -- setup/devices.sh@97 -- # nvme_mount=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:03:32.341 17:49:25 setup.sh.devices.nvme_mount -- setup/devices.sh@98 -- # nvme_dummy_test_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:03:32.341 17:49:25 setup.sh.devices.nvme_mount -- setup/devices.sh@101 -- # partition_drive nvme0n1 1 00:03:32.341 17:49:25 setup.sh.devices.nvme_mount -- setup/common.sh@39 -- # local disk=nvme0n1 00:03:32.341 17:49:25 setup.sh.devices.nvme_mount -- setup/common.sh@40 -- # local part_no=1 00:03:32.341 17:49:25 setup.sh.devices.nvme_mount -- setup/common.sh@41 -- # local size=1073741824 00:03:32.341 17:49:25 setup.sh.devices.nvme_mount -- setup/common.sh@43 -- # local part part_start=0 part_end=0 00:03:32.341 17:49:25 setup.sh.devices.nvme_mount -- setup/common.sh@44 -- # parts=() 00:03:32.341 17:49:25 setup.sh.devices.nvme_mount -- setup/common.sh@44 -- # local parts 00:03:32.341 17:49:25 setup.sh.devices.nvme_mount -- setup/common.sh@46 -- # (( part = 1 )) 00:03:32.341 17:49:25 setup.sh.devices.nvme_mount -- setup/common.sh@46 -- # (( part <= part_no )) 00:03:32.341 17:49:25 setup.sh.devices.nvme_mount -- setup/common.sh@47 -- # parts+=("${disk}p$part") 00:03:32.341 17:49:25 setup.sh.devices.nvme_mount -- setup/common.sh@46 -- # (( part++ )) 00:03:32.341 17:49:25 setup.sh.devices.nvme_mount -- setup/common.sh@46 -- # (( part <= part_no )) 00:03:32.341 17:49:25 setup.sh.devices.nvme_mount -- setup/common.sh@51 -- # (( size /= 512 )) 00:03:32.341 17:49:25 setup.sh.devices.nvme_mount -- setup/common.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/sync_dev_uevents.sh block/partition nvme0n1p1 00:03:32.341 17:49:25 setup.sh.devices.nvme_mount -- setup/common.sh@56 -- # sgdisk /dev/nvme0n1 --zap-all 00:03:32.911 Creating new GPT entries in memory. 00:03:32.911 GPT data structures destroyed! You may now partition the disk using fdisk or 00:03:32.911 other utilities. 00:03:32.911 17:49:26 setup.sh.devices.nvme_mount -- setup/common.sh@57 -- # (( part = 1 )) 00:03:32.911 17:49:26 setup.sh.devices.nvme_mount -- setup/common.sh@57 -- # (( part <= part_no )) 00:03:32.911 17:49:26 setup.sh.devices.nvme_mount -- setup/common.sh@58 -- # (( part_start = part_start == 0 ? 2048 : part_end + 1 )) 00:03:32.911 17:49:26 setup.sh.devices.nvme_mount -- setup/common.sh@59 -- # (( part_end = part_start + size - 1 )) 00:03:32.911 17:49:26 setup.sh.devices.nvme_mount -- setup/common.sh@60 -- # flock /dev/nvme0n1 sgdisk /dev/nvme0n1 --new=1:2048:2099199 00:03:34.290 Creating new GPT entries in memory. 00:03:34.290 The operation has completed successfully. 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/common.sh@57 -- # (( part++ )) 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/common.sh@57 -- # (( part <= part_no )) 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/common.sh@62 -- # wait 389344 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/devices.sh@102 -- # mkfs /dev/nvme0n1p1 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/common.sh@66 -- # local dev=/dev/nvme0n1p1 mount=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount size= 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/common.sh@68 -- # mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/common.sh@70 -- # [[ -e /dev/nvme0n1p1 ]] 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/common.sh@71 -- # mkfs.ext4 -qF /dev/nvme0n1p1 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/common.sh@72 -- # mount /dev/nvme0n1p1 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/devices.sh@105 -- # verify 0000:5e:00.0 nvme0n1:nvme0n1p1 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/devices.sh@48 -- # local dev=0000:5e:00.0 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/devices.sh@49 -- # local mounts=nvme0n1:nvme0n1p1 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/devices.sh@50 -- # local mount_point=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/devices.sh@51 -- # local test_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/devices.sh@53 -- # local found=0 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/devices.sh@55 -- # [[ -n /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme ]] 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/devices.sh@56 -- # : 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/devices.sh@59 -- # local pci status 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/devices.sh@47 -- # PCI_ALLOWED=0000:5e:00.0 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/devices.sh@47 -- # setup output config 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/common.sh@9 -- # [[ output == output ]] 00:03:34.290 17:49:27 setup.sh.devices.nvme_mount -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh config 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:5e:00.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ Active devices: mount@nvme0n1:nvme0n1p1, so not binding PCI dev == *\A\c\t\i\v\e\ \d\e\v\i\c\e\s\:\ *\n\v\m\e\0\n\1\:\n\v\m\e\0\n\1\p\1* ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@63 -- # found=1 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.7 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.6 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.4 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.3 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.2 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.1 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.7 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.6 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.4 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.3 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.2 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.1 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@66 -- # (( found == 1 )) 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@68 -- # [[ -n /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@71 -- # mountpoint -q /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@73 -- # [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@74 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@110 -- # cleanup_nvme 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@20 -- # mountpoint -q /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@21 -- # umount /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@24 -- # [[ -b /dev/nvme0n1p1 ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@25 -- # wipefs --all /dev/nvme0n1p1 00:03:36.828 /dev/nvme0n1p1: 2 bytes were erased at offset 0x00000438 (ext4): 53 ef 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@27 -- # [[ -b /dev/nvme0n1 ]] 00:03:36.828 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@28 -- # wipefs --all /dev/nvme0n1 00:03:37.088 /dev/nvme0n1: 8 bytes were erased at offset 0x00000200 (gpt): 45 46 49 20 50 41 52 54 00:03:37.088 /dev/nvme0n1: 8 bytes were erased at offset 0xe8e0db5e00 (gpt): 45 46 49 20 50 41 52 54 00:03:37.088 /dev/nvme0n1: 2 bytes were erased at offset 0x000001fe (PMBR): 55 aa 00:03:37.088 /dev/nvme0n1: calling ioctl to re-read partition table: Success 00:03:37.088 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@113 -- # mkfs /dev/nvme0n1 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 1024M 00:03:37.088 17:49:30 setup.sh.devices.nvme_mount -- setup/common.sh@66 -- # local dev=/dev/nvme0n1 mount=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount size=1024M 00:03:37.088 17:49:30 setup.sh.devices.nvme_mount -- setup/common.sh@68 -- # mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:03:37.088 17:49:30 setup.sh.devices.nvme_mount -- setup/common.sh@70 -- # [[ -e /dev/nvme0n1 ]] 00:03:37.088 17:49:30 setup.sh.devices.nvme_mount -- setup/common.sh@71 -- # mkfs.ext4 -qF /dev/nvme0n1 1024M 00:03:37.088 17:49:30 setup.sh.devices.nvme_mount -- setup/common.sh@72 -- # mount /dev/nvme0n1 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:03:37.088 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@116 -- # verify 0000:5e:00.0 nvme0n1:nvme0n1 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:03:37.088 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@48 -- # local dev=0000:5e:00.0 00:03:37.088 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@49 -- # local mounts=nvme0n1:nvme0n1 00:03:37.088 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@50 -- # local mount_point=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:03:37.088 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@51 -- # local test_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:03:37.088 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@53 -- # local found=0 00:03:37.088 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@55 -- # [[ -n /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme ]] 00:03:37.088 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@56 -- # : 00:03:37.088 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@59 -- # local pci status 00:03:37.088 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:37.088 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@47 -- # PCI_ALLOWED=0000:5e:00.0 00:03:37.088 17:49:30 setup.sh.devices.nvme_mount -- setup/devices.sh@47 -- # setup output config 00:03:37.088 17:49:30 setup.sh.devices.nvme_mount -- setup/common.sh@9 -- # [[ output == output ]] 00:03:37.088 17:49:30 setup.sh.devices.nvme_mount -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh config 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:5e:00.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ Active devices: mount@nvme0n1:nvme0n1, so not binding PCI dev == *\A\c\t\i\v\e\ \d\e\v\i\c\e\s\:\ *\n\v\m\e\0\n\1\:\n\v\m\e\0\n\1* ]] 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@63 -- # found=1 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.7 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.6 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.4 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.3 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.2 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.1 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.7 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.6 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.4 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.3 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.2 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:39.628 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:39.629 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.1 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:39.629 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:39.629 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:39.629 17:49:32 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:39.629 17:49:33 setup.sh.devices.nvme_mount -- setup/devices.sh@66 -- # (( found == 1 )) 00:03:39.629 17:49:33 setup.sh.devices.nvme_mount -- setup/devices.sh@68 -- # [[ -n /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount ]] 00:03:39.629 17:49:33 setup.sh.devices.nvme_mount -- setup/devices.sh@71 -- # mountpoint -q /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:03:39.629 17:49:33 setup.sh.devices.nvme_mount -- setup/devices.sh@73 -- # [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme ]] 00:03:39.629 17:49:33 setup.sh.devices.nvme_mount -- setup/devices.sh@74 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount/test_nvme 00:03:39.629 17:49:33 setup.sh.devices.nvme_mount -- setup/devices.sh@123 -- # umount /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:03:39.629 17:49:33 setup.sh.devices.nvme_mount -- setup/devices.sh@125 -- # verify 0000:5e:00.0 data@nvme0n1 '' '' 00:03:39.629 17:49:33 setup.sh.devices.nvme_mount -- setup/devices.sh@48 -- # local dev=0000:5e:00.0 00:03:39.629 17:49:33 setup.sh.devices.nvme_mount -- setup/devices.sh@49 -- # local mounts=data@nvme0n1 00:03:39.629 17:49:33 setup.sh.devices.nvme_mount -- setup/devices.sh@50 -- # local mount_point= 00:03:39.629 17:49:33 setup.sh.devices.nvme_mount -- setup/devices.sh@51 -- # local test_file= 00:03:39.629 17:49:33 setup.sh.devices.nvme_mount -- setup/devices.sh@53 -- # local found=0 00:03:39.629 17:49:33 setup.sh.devices.nvme_mount -- setup/devices.sh@55 -- # [[ -n '' ]] 00:03:39.629 17:49:33 setup.sh.devices.nvme_mount -- setup/devices.sh@59 -- # local pci status 00:03:39.629 17:49:33 setup.sh.devices.nvme_mount -- setup/devices.sh@47 -- # PCI_ALLOWED=0000:5e:00.0 00:03:39.629 17:49:33 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:39.629 17:49:33 setup.sh.devices.nvme_mount -- setup/devices.sh@47 -- # setup output config 00:03:39.629 17:49:33 setup.sh.devices.nvme_mount -- setup/common.sh@9 -- # [[ output == output ]] 00:03:39.629 17:49:33 setup.sh.devices.nvme_mount -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh config 00:03:42.165 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:5e:00.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ Active devices: data@nvme0n1, so not binding PCI dev == *\A\c\t\i\v\e\ \d\e\v\i\c\e\s\:\ *\d\a\t\a\@\n\v\m\e\0\n\1* ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@63 -- # found=1 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.7 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.6 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.4 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.3 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.2 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.1 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.7 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.6 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.4 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.3 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.2 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.1 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@66 -- # (( found == 1 )) 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@68 -- # [[ -n '' ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@68 -- # return 0 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@128 -- # cleanup_nvme 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@20 -- # mountpoint -q /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@24 -- # [[ -b /dev/nvme0n1p1 ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@27 -- # [[ -b /dev/nvme0n1 ]] 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- setup/devices.sh@28 -- # wipefs --all /dev/nvme0n1 00:03:42.166 /dev/nvme0n1: 2 bytes were erased at offset 0x00000438 (ext4): 53 ef 00:03:42.166 00:03:42.166 real 0m10.139s 00:03:42.166 user 0m2.906s 00:03:42.166 sys 0m5.030s 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- common/autotest_common.sh@1124 -- # xtrace_disable 00:03:42.166 17:49:35 setup.sh.devices.nvme_mount -- common/autotest_common.sh@10 -- # set +x 00:03:42.166 ************************************ 00:03:42.166 END TEST nvme_mount 00:03:42.166 ************************************ 00:03:42.166 17:49:35 setup.sh.devices -- common/autotest_common.sh@1142 -- # return 0 00:03:42.166 17:49:35 setup.sh.devices -- setup/devices.sh@214 -- # run_test dm_mount dm_mount 00:03:42.166 17:49:35 setup.sh.devices -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:03:42.166 17:49:35 setup.sh.devices -- common/autotest_common.sh@1105 -- # xtrace_disable 00:03:42.166 17:49:35 setup.sh.devices -- common/autotest_common.sh@10 -- # set +x 00:03:42.166 ************************************ 00:03:42.166 START TEST dm_mount 00:03:42.166 ************************************ 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- common/autotest_common.sh@1123 -- # dm_mount 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/devices.sh@144 -- # pv=nvme0n1 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/devices.sh@145 -- # pv0=nvme0n1p1 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/devices.sh@146 -- # pv1=nvme0n1p2 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/devices.sh@148 -- # partition_drive nvme0n1 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/common.sh@39 -- # local disk=nvme0n1 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/common.sh@40 -- # local part_no=2 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/common.sh@41 -- # local size=1073741824 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/common.sh@43 -- # local part part_start=0 part_end=0 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/common.sh@44 -- # parts=() 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/common.sh@44 -- # local parts 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/common.sh@46 -- # (( part = 1 )) 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/common.sh@46 -- # (( part <= part_no )) 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/common.sh@47 -- # parts+=("${disk}p$part") 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/common.sh@46 -- # (( part++ )) 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/common.sh@46 -- # (( part <= part_no )) 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/common.sh@47 -- # parts+=("${disk}p$part") 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/common.sh@46 -- # (( part++ )) 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/common.sh@46 -- # (( part <= part_no )) 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/common.sh@51 -- # (( size /= 512 )) 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/common.sh@56 -- # sgdisk /dev/nvme0n1 --zap-all 00:03:42.166 17:49:35 setup.sh.devices.dm_mount -- setup/common.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/sync_dev_uevents.sh block/partition nvme0n1p1 nvme0n1p2 00:03:43.105 Creating new GPT entries in memory. 00:03:43.105 GPT data structures destroyed! You may now partition the disk using fdisk or 00:03:43.105 other utilities. 00:03:43.105 17:49:36 setup.sh.devices.dm_mount -- setup/common.sh@57 -- # (( part = 1 )) 00:03:43.105 17:49:36 setup.sh.devices.dm_mount -- setup/common.sh@57 -- # (( part <= part_no )) 00:03:43.105 17:49:36 setup.sh.devices.dm_mount -- setup/common.sh@58 -- # (( part_start = part_start == 0 ? 2048 : part_end + 1 )) 00:03:43.105 17:49:36 setup.sh.devices.dm_mount -- setup/common.sh@59 -- # (( part_end = part_start + size - 1 )) 00:03:43.105 17:49:36 setup.sh.devices.dm_mount -- setup/common.sh@60 -- # flock /dev/nvme0n1 sgdisk /dev/nvme0n1 --new=1:2048:2099199 00:03:44.484 Creating new GPT entries in memory. 00:03:44.484 The operation has completed successfully. 00:03:44.484 17:49:37 setup.sh.devices.dm_mount -- setup/common.sh@57 -- # (( part++ )) 00:03:44.484 17:49:37 setup.sh.devices.dm_mount -- setup/common.sh@57 -- # (( part <= part_no )) 00:03:44.484 17:49:37 setup.sh.devices.dm_mount -- setup/common.sh@58 -- # (( part_start = part_start == 0 ? 2048 : part_end + 1 )) 00:03:44.484 17:49:37 setup.sh.devices.dm_mount -- setup/common.sh@59 -- # (( part_end = part_start + size - 1 )) 00:03:44.484 17:49:37 setup.sh.devices.dm_mount -- setup/common.sh@60 -- # flock /dev/nvme0n1 sgdisk /dev/nvme0n1 --new=2:2099200:4196351 00:03:45.423 The operation has completed successfully. 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/common.sh@57 -- # (( part++ )) 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/common.sh@57 -- # (( part <= part_no )) 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/common.sh@62 -- # wait 393302 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/devices.sh@150 -- # dm_name=nvme_dm_test 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/devices.sh@151 -- # dm_mount=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/devices.sh@152 -- # dm_dummy_test_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount/test_dm 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/devices.sh@155 -- # dmsetup create nvme_dm_test 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/devices.sh@160 -- # for t in {1..5} 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/devices.sh@161 -- # [[ -e /dev/mapper/nvme_dm_test ]] 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/devices.sh@161 -- # break 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/devices.sh@164 -- # [[ -e /dev/mapper/nvme_dm_test ]] 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/devices.sh@165 -- # readlink -f /dev/mapper/nvme_dm_test 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/devices.sh@165 -- # dm=/dev/dm-2 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/devices.sh@166 -- # dm=dm-2 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/devices.sh@168 -- # [[ -e /sys/class/block/nvme0n1p1/holders/dm-2 ]] 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/devices.sh@169 -- # [[ -e /sys/class/block/nvme0n1p2/holders/dm-2 ]] 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/devices.sh@171 -- # mkfs /dev/mapper/nvme_dm_test /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/common.sh@66 -- # local dev=/dev/mapper/nvme_dm_test mount=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount size= 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/common.sh@68 -- # mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/common.sh@70 -- # [[ -e /dev/mapper/nvme_dm_test ]] 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/common.sh@71 -- # mkfs.ext4 -qF /dev/mapper/nvme_dm_test 00:03:45.423 17:49:38 setup.sh.devices.dm_mount -- setup/common.sh@72 -- # mount /dev/mapper/nvme_dm_test /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount 00:03:45.423 17:49:39 setup.sh.devices.dm_mount -- setup/devices.sh@174 -- # verify 0000:5e:00.0 nvme0n1:nvme_dm_test /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount/test_dm 00:03:45.423 17:49:39 setup.sh.devices.dm_mount -- setup/devices.sh@48 -- # local dev=0000:5e:00.0 00:03:45.423 17:49:39 setup.sh.devices.dm_mount -- setup/devices.sh@49 -- # local mounts=nvme0n1:nvme_dm_test 00:03:45.423 17:49:39 setup.sh.devices.dm_mount -- setup/devices.sh@50 -- # local mount_point=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount 00:03:45.423 17:49:39 setup.sh.devices.dm_mount -- setup/devices.sh@51 -- # local test_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount/test_dm 00:03:45.423 17:49:39 setup.sh.devices.dm_mount -- setup/devices.sh@53 -- # local found=0 00:03:45.423 17:49:39 setup.sh.devices.dm_mount -- setup/devices.sh@55 -- # [[ -n /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount/test_dm ]] 00:03:45.423 17:49:39 setup.sh.devices.dm_mount -- setup/devices.sh@56 -- # : 00:03:45.423 17:49:39 setup.sh.devices.dm_mount -- setup/devices.sh@59 -- # local pci status 00:03:45.423 17:49:39 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:45.423 17:49:39 setup.sh.devices.dm_mount -- setup/devices.sh@47 -- # PCI_ALLOWED=0000:5e:00.0 00:03:45.423 17:49:39 setup.sh.devices.dm_mount -- setup/devices.sh@47 -- # setup output config 00:03:45.423 17:49:39 setup.sh.devices.dm_mount -- setup/common.sh@9 -- # [[ output == output ]] 00:03:45.423 17:49:39 setup.sh.devices.dm_mount -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh config 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:5e:00.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ Active devices: holder@nvme0n1p1:dm-2,holder@nvme0n1p2:dm-2,mount@nvme0n1:nvme_dm_test, so not binding PCI dev == *\A\c\t\i\v\e\ \d\e\v\i\c\e\s\:\ *\n\v\m\e\0\n\1\:\n\v\m\e\_\d\m\_\t\e\s\t* ]] 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@63 -- # found=1 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.7 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.6 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.4 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.3 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.2 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.1 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.7 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.6 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.4 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.3 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.2 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.1 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@66 -- # (( found == 1 )) 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@68 -- # [[ -n /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount ]] 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@71 -- # mountpoint -q /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount 00:03:47.962 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@73 -- # [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount/test_dm ]] 00:03:47.963 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@74 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount/test_dm 00:03:47.963 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@182 -- # umount /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount 00:03:47.963 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@184 -- # verify 0000:5e:00.0 holder@nvme0n1p1:dm-2,holder@nvme0n1p2:dm-2 '' '' 00:03:47.963 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@48 -- # local dev=0000:5e:00.0 00:03:47.963 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@49 -- # local mounts=holder@nvme0n1p1:dm-2,holder@nvme0n1p2:dm-2 00:03:47.963 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@50 -- # local mount_point= 00:03:47.963 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@51 -- # local test_file= 00:03:47.963 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@53 -- # local found=0 00:03:47.963 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@55 -- # [[ -n '' ]] 00:03:47.963 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@59 -- # local pci status 00:03:47.963 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:47.963 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@47 -- # PCI_ALLOWED=0000:5e:00.0 00:03:47.963 17:49:41 setup.sh.devices.dm_mount -- setup/devices.sh@47 -- # setup output config 00:03:47.963 17:49:41 setup.sh.devices.dm_mount -- setup/common.sh@9 -- # [[ output == output ]] 00:03:47.963 17:49:41 setup.sh.devices.dm_mount -- setup/common.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh config 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:5e:00.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ Active devices: holder@nvme0n1p1:dm-2,holder@nvme0n1p2:dm-2, so not binding PCI dev == *\A\c\t\i\v\e\ \d\e\v\i\c\e\s\:\ *\h\o\l\d\e\r\@\n\v\m\e\0\n\1\p\1\:\d\m\-\2\,\h\o\l\d\e\r\@\n\v\m\e\0\n\1\p\2\:\d\m\-\2* ]] 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@63 -- # found=1 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.7 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.6 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.4 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.3 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.2 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.1 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:00:04.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.7 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.6 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.5 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.4 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.3 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.2 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.1 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@62 -- # [[ 0000:80:04.0 == \0\0\0\0\:\5\e\:\0\0\.\0 ]] 00:03:50.499 17:49:43 setup.sh.devices.dm_mount -- setup/devices.sh@60 -- # read -r pci _ _ status 00:03:50.499 17:49:44 setup.sh.devices.dm_mount -- setup/devices.sh@66 -- # (( found == 1 )) 00:03:50.499 17:49:44 setup.sh.devices.dm_mount -- setup/devices.sh@68 -- # [[ -n '' ]] 00:03:50.499 17:49:44 setup.sh.devices.dm_mount -- setup/devices.sh@68 -- # return 0 00:03:50.499 17:49:44 setup.sh.devices.dm_mount -- setup/devices.sh@187 -- # cleanup_dm 00:03:50.499 17:49:44 setup.sh.devices.dm_mount -- setup/devices.sh@33 -- # mountpoint -q /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount 00:03:50.499 17:49:44 setup.sh.devices.dm_mount -- setup/devices.sh@36 -- # [[ -L /dev/mapper/nvme_dm_test ]] 00:03:50.499 17:49:44 setup.sh.devices.dm_mount -- setup/devices.sh@37 -- # dmsetup remove --force nvme_dm_test 00:03:50.499 17:49:44 setup.sh.devices.dm_mount -- setup/devices.sh@39 -- # [[ -b /dev/nvme0n1p1 ]] 00:03:50.499 17:49:44 setup.sh.devices.dm_mount -- setup/devices.sh@40 -- # wipefs --all /dev/nvme0n1p1 00:03:50.499 /dev/nvme0n1p1: 2 bytes were erased at offset 0x00000438 (ext4): 53 ef 00:03:50.499 17:49:44 setup.sh.devices.dm_mount -- setup/devices.sh@42 -- # [[ -b /dev/nvme0n1p2 ]] 00:03:50.499 17:49:44 setup.sh.devices.dm_mount -- setup/devices.sh@43 -- # wipefs --all /dev/nvme0n1p2 00:03:50.499 00:03:50.499 real 0m8.268s 00:03:50.499 user 0m1.895s 00:03:50.499 sys 0m3.315s 00:03:50.499 17:49:44 setup.sh.devices.dm_mount -- common/autotest_common.sh@1124 -- # xtrace_disable 00:03:50.499 17:49:44 setup.sh.devices.dm_mount -- common/autotest_common.sh@10 -- # set +x 00:03:50.499 ************************************ 00:03:50.499 END TEST dm_mount 00:03:50.499 ************************************ 00:03:50.499 17:49:44 setup.sh.devices -- common/autotest_common.sh@1142 -- # return 0 00:03:50.499 17:49:44 setup.sh.devices -- setup/devices.sh@1 -- # cleanup 00:03:50.499 17:49:44 setup.sh.devices -- setup/devices.sh@11 -- # cleanup_nvme 00:03:50.499 17:49:44 setup.sh.devices -- setup/devices.sh@20 -- # mountpoint -q /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/nvme_mount 00:03:50.499 17:49:44 setup.sh.devices -- setup/devices.sh@24 -- # [[ -b /dev/nvme0n1p1 ]] 00:03:50.499 17:49:44 setup.sh.devices -- setup/devices.sh@25 -- # wipefs --all /dev/nvme0n1p1 00:03:50.499 17:49:44 setup.sh.devices -- setup/devices.sh@27 -- # [[ -b /dev/nvme0n1 ]] 00:03:50.499 17:49:44 setup.sh.devices -- setup/devices.sh@28 -- # wipefs --all /dev/nvme0n1 00:03:50.758 /dev/nvme0n1: 8 bytes were erased at offset 0x00000200 (gpt): 45 46 49 20 50 41 52 54 00:03:50.758 /dev/nvme0n1: 8 bytes were erased at offset 0xe8e0db5e00 (gpt): 45 46 49 20 50 41 52 54 00:03:50.758 /dev/nvme0n1: 2 bytes were erased at offset 0x000001fe (PMBR): 55 aa 00:03:50.758 /dev/nvme0n1: calling ioctl to re-read partition table: Success 00:03:50.758 17:49:44 setup.sh.devices -- setup/devices.sh@12 -- # cleanup_dm 00:03:50.758 17:49:44 setup.sh.devices -- setup/devices.sh@33 -- # mountpoint -q /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/setup/dm_mount 00:03:50.758 17:49:44 setup.sh.devices -- setup/devices.sh@36 -- # [[ -L /dev/mapper/nvme_dm_test ]] 00:03:50.758 17:49:44 setup.sh.devices -- setup/devices.sh@39 -- # [[ -b /dev/nvme0n1p1 ]] 00:03:50.758 17:49:44 setup.sh.devices -- setup/devices.sh@42 -- # [[ -b /dev/nvme0n1p2 ]] 00:03:50.758 17:49:44 setup.sh.devices -- setup/devices.sh@14 -- # [[ -b /dev/nvme0n1 ]] 00:03:50.758 17:49:44 setup.sh.devices -- setup/devices.sh@15 -- # wipefs --all /dev/nvme0n1 00:03:50.758 00:03:50.758 real 0m21.518s 00:03:50.758 user 0m5.788s 00:03:50.758 sys 0m10.273s 00:03:50.758 17:49:44 setup.sh.devices -- common/autotest_common.sh@1124 -- # xtrace_disable 00:03:50.758 17:49:44 setup.sh.devices -- common/autotest_common.sh@10 -- # set +x 00:03:50.758 ************************************ 00:03:50.758 END TEST devices 00:03:50.758 ************************************ 00:03:50.758 17:49:44 setup.sh -- common/autotest_common.sh@1142 -- # return 0 00:03:50.758 00:03:50.758 real 1m12.633s 00:03:50.758 user 0m23.162s 00:03:50.758 sys 0m39.463s 00:03:50.758 17:49:44 setup.sh -- common/autotest_common.sh@1124 -- # xtrace_disable 00:03:50.758 17:49:44 setup.sh -- common/autotest_common.sh@10 -- # set +x 00:03:50.758 ************************************ 00:03:50.758 END TEST setup.sh 00:03:50.758 ************************************ 00:03:50.758 17:49:44 -- common/autotest_common.sh@1142 -- # return 0 00:03:50.758 17:49:44 -- spdk/autotest.sh@128 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh status 00:03:53.381 Hugepages 00:03:53.381 node hugesize free / total 00:03:53.381 node0 1048576kB 0 / 0 00:03:53.381 node0 2048kB 2048 / 2048 00:03:53.381 node1 1048576kB 0 / 0 00:03:53.381 node1 2048kB 0 / 0 00:03:53.381 00:03:53.381 Type BDF Vendor Device NUMA Driver Device Block devices 00:03:53.381 I/OAT 0000:00:04.0 8086 2021 0 ioatdma - - 00:03:53.381 I/OAT 0000:00:04.1 8086 2021 0 ioatdma - - 00:03:53.381 I/OAT 0000:00:04.2 8086 2021 0 ioatdma - - 00:03:53.381 I/OAT 0000:00:04.3 8086 2021 0 ioatdma - - 00:03:53.381 I/OAT 0000:00:04.4 8086 2021 0 ioatdma - - 00:03:53.381 I/OAT 0000:00:04.5 8086 2021 0 ioatdma - - 00:03:53.381 I/OAT 0000:00:04.6 8086 2021 0 ioatdma - - 00:03:53.381 I/OAT 0000:00:04.7 8086 2021 0 ioatdma - - 00:03:53.381 NVMe 0000:5e:00.0 8086 0a54 0 nvme nvme0 nvme0n1 00:03:53.381 I/OAT 0000:80:04.0 8086 2021 1 ioatdma - - 00:03:53.381 I/OAT 0000:80:04.1 8086 2021 1 ioatdma - - 00:03:53.381 I/OAT 0000:80:04.2 8086 2021 1 ioatdma - - 00:03:53.381 I/OAT 0000:80:04.3 8086 2021 1 ioatdma - - 00:03:53.381 I/OAT 0000:80:04.4 8086 2021 1 ioatdma - - 00:03:53.381 I/OAT 0000:80:04.5 8086 2021 1 ioatdma - - 00:03:53.381 I/OAT 0000:80:04.6 8086 2021 1 ioatdma - - 00:03:53.381 I/OAT 0000:80:04.7 8086 2021 1 ioatdma - - 00:03:53.381 17:49:46 -- spdk/autotest.sh@130 -- # uname -s 00:03:53.381 17:49:46 -- spdk/autotest.sh@130 -- # [[ Linux == Linux ]] 00:03:53.381 17:49:46 -- spdk/autotest.sh@132 -- # nvme_namespace_revert 00:03:53.381 17:49:46 -- common/autotest_common.sh@1531 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:03:55.917 0000:00:04.7 (8086 2021): ioatdma -> vfio-pci 00:03:55.917 0000:00:04.6 (8086 2021): ioatdma -> vfio-pci 00:03:55.917 0000:00:04.5 (8086 2021): ioatdma -> vfio-pci 00:03:55.917 0000:00:04.4 (8086 2021): ioatdma -> vfio-pci 00:03:55.917 0000:00:04.3 (8086 2021): ioatdma -> vfio-pci 00:03:55.917 0000:00:04.2 (8086 2021): ioatdma -> vfio-pci 00:03:55.917 0000:00:04.1 (8086 2021): ioatdma -> vfio-pci 00:03:55.917 0000:00:04.0 (8086 2021): ioatdma -> vfio-pci 00:03:55.917 0000:80:04.7 (8086 2021): ioatdma -> vfio-pci 00:03:55.917 0000:80:04.6 (8086 2021): ioatdma -> vfio-pci 00:03:55.917 0000:80:04.5 (8086 2021): ioatdma -> vfio-pci 00:03:55.917 0000:80:04.4 (8086 2021): ioatdma -> vfio-pci 00:03:55.917 0000:80:04.3 (8086 2021): ioatdma -> vfio-pci 00:03:55.917 0000:80:04.2 (8086 2021): ioatdma -> vfio-pci 00:03:55.917 0000:80:04.1 (8086 2021): ioatdma -> vfio-pci 00:03:55.917 0000:80:04.0 (8086 2021): ioatdma -> vfio-pci 00:03:56.485 0000:5e:00.0 (8086 0a54): nvme -> vfio-pci 00:03:56.744 17:49:50 -- common/autotest_common.sh@1532 -- # sleep 1 00:03:57.681 17:49:51 -- common/autotest_common.sh@1533 -- # bdfs=() 00:03:57.681 17:49:51 -- common/autotest_common.sh@1533 -- # local bdfs 00:03:57.681 17:49:51 -- common/autotest_common.sh@1534 -- # bdfs=($(get_nvme_bdfs)) 00:03:57.681 17:49:51 -- common/autotest_common.sh@1534 -- # get_nvme_bdfs 00:03:57.681 17:49:51 -- common/autotest_common.sh@1513 -- # bdfs=() 00:03:57.681 17:49:51 -- common/autotest_common.sh@1513 -- # local bdfs 00:03:57.681 17:49:51 -- common/autotest_common.sh@1514 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:03:57.681 17:49:51 -- common/autotest_common.sh@1514 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:03:57.681 17:49:51 -- common/autotest_common.sh@1514 -- # jq -r '.config[].params.traddr' 00:03:57.681 17:49:51 -- common/autotest_common.sh@1515 -- # (( 1 == 0 )) 00:03:57.681 17:49:51 -- common/autotest_common.sh@1519 -- # printf '%s\n' 0000:5e:00.0 00:03:57.681 17:49:51 -- common/autotest_common.sh@1536 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:04:00.215 Waiting for block devices as requested 00:04:00.215 0000:5e:00.0 (8086 0a54): vfio-pci -> nvme 00:04:00.215 0000:00:04.7 (8086 2021): vfio-pci -> ioatdma 00:04:00.215 0000:00:04.6 (8086 2021): vfio-pci -> ioatdma 00:04:00.215 0000:00:04.5 (8086 2021): vfio-pci -> ioatdma 00:04:00.473 0000:00:04.4 (8086 2021): vfio-pci -> ioatdma 00:04:00.473 0000:00:04.3 (8086 2021): vfio-pci -> ioatdma 00:04:00.473 0000:00:04.2 (8086 2021): vfio-pci -> ioatdma 00:04:00.473 0000:00:04.1 (8086 2021): vfio-pci -> ioatdma 00:04:00.732 0000:00:04.0 (8086 2021): vfio-pci -> ioatdma 00:04:00.732 0000:80:04.7 (8086 2021): vfio-pci -> ioatdma 00:04:00.732 0000:80:04.6 (8086 2021): vfio-pci -> ioatdma 00:04:00.992 0000:80:04.5 (8086 2021): vfio-pci -> ioatdma 00:04:00.992 0000:80:04.4 (8086 2021): vfio-pci -> ioatdma 00:04:00.992 0000:80:04.3 (8086 2021): vfio-pci -> ioatdma 00:04:00.992 0000:80:04.2 (8086 2021): vfio-pci -> ioatdma 00:04:01.251 0000:80:04.1 (8086 2021): vfio-pci -> ioatdma 00:04:01.251 0000:80:04.0 (8086 2021): vfio-pci -> ioatdma 00:04:01.251 17:49:54 -- common/autotest_common.sh@1538 -- # for bdf in "${bdfs[@]}" 00:04:01.251 17:49:54 -- common/autotest_common.sh@1539 -- # get_nvme_ctrlr_from_bdf 0000:5e:00.0 00:04:01.251 17:49:54 -- common/autotest_common.sh@1502 -- # readlink -f /sys/class/nvme/nvme0 00:04:01.251 17:49:54 -- common/autotest_common.sh@1502 -- # grep 0000:5e:00.0/nvme/nvme 00:04:01.251 17:49:54 -- common/autotest_common.sh@1502 -- # bdf_sysfs_path=/sys/devices/pci0000:5d/0000:5d:02.0/0000:5e:00.0/nvme/nvme0 00:04:01.251 17:49:54 -- common/autotest_common.sh@1503 -- # [[ -z /sys/devices/pci0000:5d/0000:5d:02.0/0000:5e:00.0/nvme/nvme0 ]] 00:04:01.251 17:49:54 -- common/autotest_common.sh@1507 -- # basename /sys/devices/pci0000:5d/0000:5d:02.0/0000:5e:00.0/nvme/nvme0 00:04:01.251 17:49:54 -- common/autotest_common.sh@1507 -- # printf '%s\n' nvme0 00:04:01.251 17:49:54 -- common/autotest_common.sh@1539 -- # nvme_ctrlr=/dev/nvme0 00:04:01.251 17:49:54 -- common/autotest_common.sh@1540 -- # [[ -z /dev/nvme0 ]] 00:04:01.251 17:49:54 -- common/autotest_common.sh@1545 -- # nvme id-ctrl /dev/nvme0 00:04:01.251 17:49:54 -- common/autotest_common.sh@1545 -- # grep oacs 00:04:01.251 17:49:54 -- common/autotest_common.sh@1545 -- # cut -d: -f2 00:04:01.251 17:49:54 -- common/autotest_common.sh@1545 -- # oacs=' 0xe' 00:04:01.251 17:49:54 -- common/autotest_common.sh@1546 -- # oacs_ns_manage=8 00:04:01.251 17:49:54 -- common/autotest_common.sh@1548 -- # [[ 8 -ne 0 ]] 00:04:01.251 17:49:54 -- common/autotest_common.sh@1554 -- # nvme id-ctrl /dev/nvme0 00:04:01.251 17:49:54 -- common/autotest_common.sh@1554 -- # grep unvmcap 00:04:01.251 17:49:54 -- common/autotest_common.sh@1554 -- # cut -d: -f2 00:04:01.251 17:49:54 -- common/autotest_common.sh@1554 -- # unvmcap=' 0' 00:04:01.251 17:49:54 -- common/autotest_common.sh@1555 -- # [[ 0 -eq 0 ]] 00:04:01.251 17:49:54 -- common/autotest_common.sh@1557 -- # continue 00:04:01.251 17:49:54 -- spdk/autotest.sh@135 -- # timing_exit pre_cleanup 00:04:01.251 17:49:54 -- common/autotest_common.sh@728 -- # xtrace_disable 00:04:01.251 17:49:54 -- common/autotest_common.sh@10 -- # set +x 00:04:01.510 17:49:54 -- spdk/autotest.sh@138 -- # timing_enter afterboot 00:04:01.510 17:49:54 -- common/autotest_common.sh@722 -- # xtrace_disable 00:04:01.510 17:49:54 -- common/autotest_common.sh@10 -- # set +x 00:04:01.510 17:49:55 -- spdk/autotest.sh@139 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:04:04.047 0000:00:04.7 (8086 2021): ioatdma -> vfio-pci 00:04:04.047 0000:00:04.6 (8086 2021): ioatdma -> vfio-pci 00:04:04.047 0000:00:04.5 (8086 2021): ioatdma -> vfio-pci 00:04:04.047 0000:00:04.4 (8086 2021): ioatdma -> vfio-pci 00:04:04.047 0000:00:04.3 (8086 2021): ioatdma -> vfio-pci 00:04:04.047 0000:00:04.2 (8086 2021): ioatdma -> vfio-pci 00:04:04.047 0000:00:04.1 (8086 2021): ioatdma -> vfio-pci 00:04:04.047 0000:00:04.0 (8086 2021): ioatdma -> vfio-pci 00:04:04.047 0000:80:04.7 (8086 2021): ioatdma -> vfio-pci 00:04:04.047 0000:80:04.6 (8086 2021): ioatdma -> vfio-pci 00:04:04.047 0000:80:04.5 (8086 2021): ioatdma -> vfio-pci 00:04:04.047 0000:80:04.4 (8086 2021): ioatdma -> vfio-pci 00:04:04.047 0000:80:04.3 (8086 2021): ioatdma -> vfio-pci 00:04:04.047 0000:80:04.2 (8086 2021): ioatdma -> vfio-pci 00:04:04.047 0000:80:04.1 (8086 2021): ioatdma -> vfio-pci 00:04:04.047 0000:80:04.0 (8086 2021): ioatdma -> vfio-pci 00:04:04.988 0000:5e:00.0 (8086 0a54): nvme -> vfio-pci 00:04:04.988 17:49:58 -- spdk/autotest.sh@140 -- # timing_exit afterboot 00:04:04.988 17:49:58 -- common/autotest_common.sh@728 -- # xtrace_disable 00:04:04.988 17:49:58 -- common/autotest_common.sh@10 -- # set +x 00:04:04.988 17:49:58 -- spdk/autotest.sh@144 -- # opal_revert_cleanup 00:04:04.988 17:49:58 -- common/autotest_common.sh@1591 -- # mapfile -t bdfs 00:04:04.988 17:49:58 -- common/autotest_common.sh@1591 -- # get_nvme_bdfs_by_id 0x0a54 00:04:04.988 17:49:58 -- common/autotest_common.sh@1577 -- # bdfs=() 00:04:04.988 17:49:58 -- common/autotest_common.sh@1577 -- # local bdfs 00:04:04.988 17:49:58 -- common/autotest_common.sh@1579 -- # get_nvme_bdfs 00:04:04.988 17:49:58 -- common/autotest_common.sh@1513 -- # bdfs=() 00:04:04.988 17:49:58 -- common/autotest_common.sh@1513 -- # local bdfs 00:04:04.988 17:49:58 -- common/autotest_common.sh@1514 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:04:04.988 17:49:58 -- common/autotest_common.sh@1514 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:04:04.988 17:49:58 -- common/autotest_common.sh@1514 -- # jq -r '.config[].params.traddr' 00:04:04.988 17:49:58 -- common/autotest_common.sh@1515 -- # (( 1 == 0 )) 00:04:04.988 17:49:58 -- common/autotest_common.sh@1519 -- # printf '%s\n' 0000:5e:00.0 00:04:04.988 17:49:58 -- common/autotest_common.sh@1579 -- # for bdf in $(get_nvme_bdfs) 00:04:04.988 17:49:58 -- common/autotest_common.sh@1580 -- # cat /sys/bus/pci/devices/0000:5e:00.0/device 00:04:04.988 17:49:58 -- common/autotest_common.sh@1580 -- # device=0x0a54 00:04:04.988 17:49:58 -- common/autotest_common.sh@1581 -- # [[ 0x0a54 == \0\x\0\a\5\4 ]] 00:04:04.988 17:49:58 -- common/autotest_common.sh@1582 -- # bdfs+=($bdf) 00:04:04.988 17:49:58 -- common/autotest_common.sh@1586 -- # printf '%s\n' 0000:5e:00.0 00:04:04.988 17:49:58 -- common/autotest_common.sh@1592 -- # [[ -z 0000:5e:00.0 ]] 00:04:04.988 17:49:58 -- common/autotest_common.sh@1597 -- # spdk_tgt_pid=402072 00:04:04.988 17:49:58 -- common/autotest_common.sh@1596 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:04.988 17:49:58 -- common/autotest_common.sh@1598 -- # waitforlisten 402072 00:04:04.988 17:49:58 -- common/autotest_common.sh@829 -- # '[' -z 402072 ']' 00:04:04.988 17:49:58 -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:04.988 17:49:58 -- common/autotest_common.sh@834 -- # local max_retries=100 00:04:04.988 17:49:58 -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:04.988 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:04.988 17:49:58 -- common/autotest_common.sh@838 -- # xtrace_disable 00:04:04.988 17:49:58 -- common/autotest_common.sh@10 -- # set +x 00:04:05.247 [2024-07-15 17:49:58.732095] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:04:05.247 [2024-07-15 17:49:58.732142] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid402072 ] 00:04:05.247 EAL: No free 2048 kB hugepages reported on node 1 00:04:05.247 [2024-07-15 17:49:58.785620] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:05.247 [2024-07-15 17:49:58.867408] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:04:05.816 17:49:59 -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:04:05.816 17:49:59 -- common/autotest_common.sh@862 -- # return 0 00:04:05.816 17:49:59 -- common/autotest_common.sh@1600 -- # bdf_id=0 00:04:05.816 17:49:59 -- common/autotest_common.sh@1601 -- # for bdf in "${bdfs[@]}" 00:04:05.816 17:49:59 -- common/autotest_common.sh@1602 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_nvme_attach_controller -b nvme0 -t pcie -a 0000:5e:00.0 00:04:09.106 nvme0n1 00:04:09.106 17:50:02 -- common/autotest_common.sh@1604 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_nvme_opal_revert -b nvme0 -p test 00:04:09.106 [2024-07-15 17:50:02.687899] vbdev_opal_rpc.c: 125:rpc_bdev_nvme_opal_revert: *ERROR*: nvme0 not support opal 00:04:09.106 request: 00:04:09.106 { 00:04:09.106 "nvme_ctrlr_name": "nvme0", 00:04:09.106 "password": "test", 00:04:09.106 "method": "bdev_nvme_opal_revert", 00:04:09.106 "req_id": 1 00:04:09.106 } 00:04:09.106 Got JSON-RPC error response 00:04:09.106 response: 00:04:09.106 { 00:04:09.106 "code": -32602, 00:04:09.106 "message": "Invalid parameters" 00:04:09.106 } 00:04:09.106 17:50:02 -- common/autotest_common.sh@1604 -- # true 00:04:09.106 17:50:02 -- common/autotest_common.sh@1605 -- # (( ++bdf_id )) 00:04:09.106 17:50:02 -- common/autotest_common.sh@1608 -- # killprocess 402072 00:04:09.106 17:50:02 -- common/autotest_common.sh@948 -- # '[' -z 402072 ']' 00:04:09.106 17:50:02 -- common/autotest_common.sh@952 -- # kill -0 402072 00:04:09.106 17:50:02 -- common/autotest_common.sh@953 -- # uname 00:04:09.106 17:50:02 -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:04:09.106 17:50:02 -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 402072 00:04:09.106 17:50:02 -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:04:09.106 17:50:02 -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:04:09.106 17:50:02 -- common/autotest_common.sh@966 -- # echo 'killing process with pid 402072' 00:04:09.106 killing process with pid 402072 00:04:09.106 17:50:02 -- common/autotest_common.sh@967 -- # kill 402072 00:04:09.106 17:50:02 -- common/autotest_common.sh@972 -- # wait 402072 00:04:11.007 17:50:04 -- spdk/autotest.sh@150 -- # '[' 0 -eq 1 ']' 00:04:11.007 17:50:04 -- spdk/autotest.sh@154 -- # '[' 1 -eq 1 ']' 00:04:11.007 17:50:04 -- spdk/autotest.sh@155 -- # [[ 0 -eq 1 ]] 00:04:11.007 17:50:04 -- spdk/autotest.sh@155 -- # [[ 0 -eq 1 ]] 00:04:11.007 17:50:04 -- spdk/autotest.sh@162 -- # timing_enter lib 00:04:11.007 17:50:04 -- common/autotest_common.sh@722 -- # xtrace_disable 00:04:11.007 17:50:04 -- common/autotest_common.sh@10 -- # set +x 00:04:11.007 17:50:04 -- spdk/autotest.sh@164 -- # [[ 0 -eq 1 ]] 00:04:11.007 17:50:04 -- spdk/autotest.sh@168 -- # run_test env /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env.sh 00:04:11.007 17:50:04 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:11.007 17:50:04 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:11.007 17:50:04 -- common/autotest_common.sh@10 -- # set +x 00:04:11.007 ************************************ 00:04:11.007 START TEST env 00:04:11.007 ************************************ 00:04:11.007 17:50:04 env -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env.sh 00:04:11.007 * Looking for test storage... 00:04:11.007 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env 00:04:11.007 17:50:04 env -- env/env.sh@10 -- # run_test env_memory /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/memory/memory_ut 00:04:11.007 17:50:04 env -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:11.007 17:50:04 env -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:11.007 17:50:04 env -- common/autotest_common.sh@10 -- # set +x 00:04:11.007 ************************************ 00:04:11.007 START TEST env_memory 00:04:11.007 ************************************ 00:04:11.007 17:50:04 env.env_memory -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/memory/memory_ut 00:04:11.007 00:04:11.007 00:04:11.007 CUnit - A unit testing framework for C - Version 2.1-3 00:04:11.007 http://cunit.sourceforge.net/ 00:04:11.007 00:04:11.007 00:04:11.007 Suite: memory 00:04:11.007 Test: alloc and free memory map ...[2024-07-15 17:50:04.525096] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 283:spdk_mem_map_alloc: *ERROR*: Initial mem_map notify failed 00:04:11.007 passed 00:04:11.007 Test: mem map translation ...[2024-07-15 17:50:04.543019] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 590:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=2097152 len=1234 00:04:11.007 [2024-07-15 17:50:04.543034] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 590:spdk_mem_map_set_translation: *ERROR*: invalid spdk_mem_map_set_translation parameters, vaddr=1234 len=2097152 00:04:11.007 [2024-07-15 17:50:04.543068] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 584:spdk_mem_map_set_translation: *ERROR*: invalid usermode virtual address 281474976710656 00:04:11.007 [2024-07-15 17:50:04.543076] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 600:spdk_mem_map_set_translation: *ERROR*: could not get 0xffffffe00000 map 00:04:11.007 passed 00:04:11.007 Test: mem map registration ...[2024-07-15 17:50:04.579613] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 346:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=0x200000 len=1234 00:04:11.007 [2024-07-15 17:50:04.579631] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/memory.c: 346:spdk_mem_register: *ERROR*: invalid spdk_mem_register parameters, vaddr=0x4d2 len=2097152 00:04:11.007 passed 00:04:11.007 Test: mem map adjacent registrations ...passed 00:04:11.007 00:04:11.007 Run Summary: Type Total Ran Passed Failed Inactive 00:04:11.007 suites 1 1 n/a 0 0 00:04:11.007 tests 4 4 4 0 0 00:04:11.007 asserts 152 152 152 0 n/a 00:04:11.007 00:04:11.007 Elapsed time = 0.128 seconds 00:04:11.007 00:04:11.007 real 0m0.135s 00:04:11.007 user 0m0.128s 00:04:11.007 sys 0m0.007s 00:04:11.007 17:50:04 env.env_memory -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:11.007 17:50:04 env.env_memory -- common/autotest_common.sh@10 -- # set +x 00:04:11.007 ************************************ 00:04:11.007 END TEST env_memory 00:04:11.007 ************************************ 00:04:11.007 17:50:04 env -- common/autotest_common.sh@1142 -- # return 0 00:04:11.007 17:50:04 env -- env/env.sh@11 -- # run_test env_vtophys /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/vtophys/vtophys 00:04:11.007 17:50:04 env -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:11.007 17:50:04 env -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:11.007 17:50:04 env -- common/autotest_common.sh@10 -- # set +x 00:04:11.007 ************************************ 00:04:11.007 START TEST env_vtophys 00:04:11.007 ************************************ 00:04:11.007 17:50:04 env.env_vtophys -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/vtophys/vtophys 00:04:11.007 EAL: lib.eal log level changed from notice to debug 00:04:11.007 EAL: Detected lcore 0 as core 0 on socket 0 00:04:11.007 EAL: Detected lcore 1 as core 1 on socket 0 00:04:11.007 EAL: Detected lcore 2 as core 2 on socket 0 00:04:11.007 EAL: Detected lcore 3 as core 3 on socket 0 00:04:11.007 EAL: Detected lcore 4 as core 4 on socket 0 00:04:11.007 EAL: Detected lcore 5 as core 5 on socket 0 00:04:11.007 EAL: Detected lcore 6 as core 6 on socket 0 00:04:11.007 EAL: Detected lcore 7 as core 8 on socket 0 00:04:11.008 EAL: Detected lcore 8 as core 9 on socket 0 00:04:11.008 EAL: Detected lcore 9 as core 10 on socket 0 00:04:11.008 EAL: Detected lcore 10 as core 11 on socket 0 00:04:11.008 EAL: Detected lcore 11 as core 12 on socket 0 00:04:11.008 EAL: Detected lcore 12 as core 13 on socket 0 00:04:11.008 EAL: Detected lcore 13 as core 16 on socket 0 00:04:11.008 EAL: Detected lcore 14 as core 17 on socket 0 00:04:11.008 EAL: Detected lcore 15 as core 18 on socket 0 00:04:11.008 EAL: Detected lcore 16 as core 19 on socket 0 00:04:11.008 EAL: Detected lcore 17 as core 20 on socket 0 00:04:11.008 EAL: Detected lcore 18 as core 21 on socket 0 00:04:11.008 EAL: Detected lcore 19 as core 25 on socket 0 00:04:11.008 EAL: Detected lcore 20 as core 26 on socket 0 00:04:11.008 EAL: Detected lcore 21 as core 27 on socket 0 00:04:11.008 EAL: Detected lcore 22 as core 28 on socket 0 00:04:11.008 EAL: Detected lcore 23 as core 29 on socket 0 00:04:11.008 EAL: Detected lcore 24 as core 0 on socket 1 00:04:11.008 EAL: Detected lcore 25 as core 1 on socket 1 00:04:11.008 EAL: Detected lcore 26 as core 2 on socket 1 00:04:11.008 EAL: Detected lcore 27 as core 3 on socket 1 00:04:11.008 EAL: Detected lcore 28 as core 4 on socket 1 00:04:11.008 EAL: Detected lcore 29 as core 5 on socket 1 00:04:11.008 EAL: Detected lcore 30 as core 6 on socket 1 00:04:11.008 EAL: Detected lcore 31 as core 9 on socket 1 00:04:11.008 EAL: Detected lcore 32 as core 10 on socket 1 00:04:11.008 EAL: Detected lcore 33 as core 11 on socket 1 00:04:11.008 EAL: Detected lcore 34 as core 12 on socket 1 00:04:11.008 EAL: Detected lcore 35 as core 13 on socket 1 00:04:11.008 EAL: Detected lcore 36 as core 16 on socket 1 00:04:11.008 EAL: Detected lcore 37 as core 17 on socket 1 00:04:11.008 EAL: Detected lcore 38 as core 18 on socket 1 00:04:11.008 EAL: Detected lcore 39 as core 19 on socket 1 00:04:11.008 EAL: Detected lcore 40 as core 20 on socket 1 00:04:11.008 EAL: Detected lcore 41 as core 21 on socket 1 00:04:11.008 EAL: Detected lcore 42 as core 24 on socket 1 00:04:11.008 EAL: Detected lcore 43 as core 25 on socket 1 00:04:11.008 EAL: Detected lcore 44 as core 26 on socket 1 00:04:11.008 EAL: Detected lcore 45 as core 27 on socket 1 00:04:11.008 EAL: Detected lcore 46 as core 28 on socket 1 00:04:11.008 EAL: Detected lcore 47 as core 29 on socket 1 00:04:11.008 EAL: Detected lcore 48 as core 0 on socket 0 00:04:11.008 EAL: Detected lcore 49 as core 1 on socket 0 00:04:11.008 EAL: Detected lcore 50 as core 2 on socket 0 00:04:11.008 EAL: Detected lcore 51 as core 3 on socket 0 00:04:11.008 EAL: Detected lcore 52 as core 4 on socket 0 00:04:11.008 EAL: Detected lcore 53 as core 5 on socket 0 00:04:11.008 EAL: Detected lcore 54 as core 6 on socket 0 00:04:11.008 EAL: Detected lcore 55 as core 8 on socket 0 00:04:11.008 EAL: Detected lcore 56 as core 9 on socket 0 00:04:11.008 EAL: Detected lcore 57 as core 10 on socket 0 00:04:11.008 EAL: Detected lcore 58 as core 11 on socket 0 00:04:11.008 EAL: Detected lcore 59 as core 12 on socket 0 00:04:11.008 EAL: Detected lcore 60 as core 13 on socket 0 00:04:11.008 EAL: Detected lcore 61 as core 16 on socket 0 00:04:11.008 EAL: Detected lcore 62 as core 17 on socket 0 00:04:11.008 EAL: Detected lcore 63 as core 18 on socket 0 00:04:11.008 EAL: Detected lcore 64 as core 19 on socket 0 00:04:11.008 EAL: Detected lcore 65 as core 20 on socket 0 00:04:11.008 EAL: Detected lcore 66 as core 21 on socket 0 00:04:11.008 EAL: Detected lcore 67 as core 25 on socket 0 00:04:11.008 EAL: Detected lcore 68 as core 26 on socket 0 00:04:11.008 EAL: Detected lcore 69 as core 27 on socket 0 00:04:11.008 EAL: Detected lcore 70 as core 28 on socket 0 00:04:11.008 EAL: Detected lcore 71 as core 29 on socket 0 00:04:11.008 EAL: Detected lcore 72 as core 0 on socket 1 00:04:11.008 EAL: Detected lcore 73 as core 1 on socket 1 00:04:11.008 EAL: Detected lcore 74 as core 2 on socket 1 00:04:11.008 EAL: Detected lcore 75 as core 3 on socket 1 00:04:11.008 EAL: Detected lcore 76 as core 4 on socket 1 00:04:11.008 EAL: Detected lcore 77 as core 5 on socket 1 00:04:11.008 EAL: Detected lcore 78 as core 6 on socket 1 00:04:11.008 EAL: Detected lcore 79 as core 9 on socket 1 00:04:11.008 EAL: Detected lcore 80 as core 10 on socket 1 00:04:11.008 EAL: Detected lcore 81 as core 11 on socket 1 00:04:11.008 EAL: Detected lcore 82 as core 12 on socket 1 00:04:11.008 EAL: Detected lcore 83 as core 13 on socket 1 00:04:11.008 EAL: Detected lcore 84 as core 16 on socket 1 00:04:11.008 EAL: Detected lcore 85 as core 17 on socket 1 00:04:11.008 EAL: Detected lcore 86 as core 18 on socket 1 00:04:11.008 EAL: Detected lcore 87 as core 19 on socket 1 00:04:11.008 EAL: Detected lcore 88 as core 20 on socket 1 00:04:11.008 EAL: Detected lcore 89 as core 21 on socket 1 00:04:11.008 EAL: Detected lcore 90 as core 24 on socket 1 00:04:11.008 EAL: Detected lcore 91 as core 25 on socket 1 00:04:11.008 EAL: Detected lcore 92 as core 26 on socket 1 00:04:11.008 EAL: Detected lcore 93 as core 27 on socket 1 00:04:11.008 EAL: Detected lcore 94 as core 28 on socket 1 00:04:11.008 EAL: Detected lcore 95 as core 29 on socket 1 00:04:11.008 EAL: Maximum logical cores by configuration: 128 00:04:11.008 EAL: Detected CPU lcores: 96 00:04:11.008 EAL: Detected NUMA nodes: 2 00:04:11.008 EAL: Checking presence of .so 'librte_eal.so.24.1' 00:04:11.008 EAL: Detected shared linkage of DPDK 00:04:11.008 EAL: No shared files mode enabled, IPC will be disabled 00:04:11.267 EAL: Bus pci wants IOVA as 'DC' 00:04:11.267 EAL: Buses did not request a specific IOVA mode. 00:04:11.267 EAL: IOMMU is available, selecting IOVA as VA mode. 00:04:11.267 EAL: Selected IOVA mode 'VA' 00:04:11.267 EAL: No free 2048 kB hugepages reported on node 1 00:04:11.267 EAL: Probing VFIO support... 00:04:11.267 EAL: IOMMU type 1 (Type 1) is supported 00:04:11.267 EAL: IOMMU type 7 (sPAPR) is not supported 00:04:11.267 EAL: IOMMU type 8 (No-IOMMU) is not supported 00:04:11.267 EAL: VFIO support initialized 00:04:11.267 EAL: Ask a virtual area of 0x2e000 bytes 00:04:11.267 EAL: Virtual area found at 0x200000000000 (size = 0x2e000) 00:04:11.267 EAL: Setting up physically contiguous memory... 00:04:11.267 EAL: Setting maximum number of open files to 524288 00:04:11.267 EAL: Detected memory type: socket_id:0 hugepage_sz:2097152 00:04:11.267 EAL: Detected memory type: socket_id:1 hugepage_sz:2097152 00:04:11.267 EAL: Creating 4 segment lists: n_segs:8192 socket_id:0 hugepage_sz:2097152 00:04:11.267 EAL: Ask a virtual area of 0x61000 bytes 00:04:11.267 EAL: Virtual area found at 0x20000002e000 (size = 0x61000) 00:04:11.267 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:04:11.267 EAL: Ask a virtual area of 0x400000000 bytes 00:04:11.267 EAL: Virtual area found at 0x200000200000 (size = 0x400000000) 00:04:11.267 EAL: VA reserved for memseg list at 0x200000200000, size 400000000 00:04:11.267 EAL: Ask a virtual area of 0x61000 bytes 00:04:11.267 EAL: Virtual area found at 0x200400200000 (size = 0x61000) 00:04:11.267 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:04:11.267 EAL: Ask a virtual area of 0x400000000 bytes 00:04:11.267 EAL: Virtual area found at 0x200400400000 (size = 0x400000000) 00:04:11.267 EAL: VA reserved for memseg list at 0x200400400000, size 400000000 00:04:11.267 EAL: Ask a virtual area of 0x61000 bytes 00:04:11.267 EAL: Virtual area found at 0x200800400000 (size = 0x61000) 00:04:11.267 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:04:11.267 EAL: Ask a virtual area of 0x400000000 bytes 00:04:11.267 EAL: Virtual area found at 0x200800600000 (size = 0x400000000) 00:04:11.267 EAL: VA reserved for memseg list at 0x200800600000, size 400000000 00:04:11.267 EAL: Ask a virtual area of 0x61000 bytes 00:04:11.267 EAL: Virtual area found at 0x200c00600000 (size = 0x61000) 00:04:11.267 EAL: Memseg list allocated at socket 0, page size 0x800kB 00:04:11.267 EAL: Ask a virtual area of 0x400000000 bytes 00:04:11.267 EAL: Virtual area found at 0x200c00800000 (size = 0x400000000) 00:04:11.267 EAL: VA reserved for memseg list at 0x200c00800000, size 400000000 00:04:11.267 EAL: Creating 4 segment lists: n_segs:8192 socket_id:1 hugepage_sz:2097152 00:04:11.267 EAL: Ask a virtual area of 0x61000 bytes 00:04:11.267 EAL: Virtual area found at 0x201000800000 (size = 0x61000) 00:04:11.267 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:04:11.267 EAL: Ask a virtual area of 0x400000000 bytes 00:04:11.267 EAL: Virtual area found at 0x201000a00000 (size = 0x400000000) 00:04:11.267 EAL: VA reserved for memseg list at 0x201000a00000, size 400000000 00:04:11.267 EAL: Ask a virtual area of 0x61000 bytes 00:04:11.267 EAL: Virtual area found at 0x201400a00000 (size = 0x61000) 00:04:11.267 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:04:11.267 EAL: Ask a virtual area of 0x400000000 bytes 00:04:11.267 EAL: Virtual area found at 0x201400c00000 (size = 0x400000000) 00:04:11.267 EAL: VA reserved for memseg list at 0x201400c00000, size 400000000 00:04:11.267 EAL: Ask a virtual area of 0x61000 bytes 00:04:11.267 EAL: Virtual area found at 0x201800c00000 (size = 0x61000) 00:04:11.267 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:04:11.267 EAL: Ask a virtual area of 0x400000000 bytes 00:04:11.267 EAL: Virtual area found at 0x201800e00000 (size = 0x400000000) 00:04:11.267 EAL: VA reserved for memseg list at 0x201800e00000, size 400000000 00:04:11.267 EAL: Ask a virtual area of 0x61000 bytes 00:04:11.267 EAL: Virtual area found at 0x201c00e00000 (size = 0x61000) 00:04:11.267 EAL: Memseg list allocated at socket 1, page size 0x800kB 00:04:11.267 EAL: Ask a virtual area of 0x400000000 bytes 00:04:11.267 EAL: Virtual area found at 0x201c01000000 (size = 0x400000000) 00:04:11.267 EAL: VA reserved for memseg list at 0x201c01000000, size 400000000 00:04:11.267 EAL: Hugepages will be freed exactly as allocated. 00:04:11.267 EAL: No shared files mode enabled, IPC is disabled 00:04:11.267 EAL: No shared files mode enabled, IPC is disabled 00:04:11.267 EAL: TSC frequency is ~2300000 KHz 00:04:11.267 EAL: Main lcore 0 is ready (tid=7f5864573a00;cpuset=[0]) 00:04:11.267 EAL: Trying to obtain current memory policy. 00:04:11.267 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:11.267 EAL: Restoring previous memory policy: 0 00:04:11.267 EAL: request: mp_malloc_sync 00:04:11.267 EAL: No shared files mode enabled, IPC is disabled 00:04:11.267 EAL: Heap on socket 0 was expanded by 2MB 00:04:11.267 EAL: No shared files mode enabled, IPC is disabled 00:04:11.267 EAL: No PCI address specified using 'addr=' in: bus=pci 00:04:11.267 EAL: Mem event callback 'spdk:(nil)' registered 00:04:11.267 00:04:11.267 00:04:11.267 CUnit - A unit testing framework for C - Version 2.1-3 00:04:11.267 http://cunit.sourceforge.net/ 00:04:11.267 00:04:11.267 00:04:11.267 Suite: components_suite 00:04:11.267 Test: vtophys_malloc_test ...passed 00:04:11.267 Test: vtophys_spdk_malloc_test ...EAL: Trying to obtain current memory policy. 00:04:11.267 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:11.267 EAL: Restoring previous memory policy: 4 00:04:11.267 EAL: Calling mem event callback 'spdk:(nil)' 00:04:11.267 EAL: request: mp_malloc_sync 00:04:11.267 EAL: No shared files mode enabled, IPC is disabled 00:04:11.267 EAL: Heap on socket 0 was expanded by 4MB 00:04:11.267 EAL: Calling mem event callback 'spdk:(nil)' 00:04:11.267 EAL: request: mp_malloc_sync 00:04:11.267 EAL: No shared files mode enabled, IPC is disabled 00:04:11.267 EAL: Heap on socket 0 was shrunk by 4MB 00:04:11.267 EAL: Trying to obtain current memory policy. 00:04:11.267 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:11.267 EAL: Restoring previous memory policy: 4 00:04:11.267 EAL: Calling mem event callback 'spdk:(nil)' 00:04:11.267 EAL: request: mp_malloc_sync 00:04:11.267 EAL: No shared files mode enabled, IPC is disabled 00:04:11.267 EAL: Heap on socket 0 was expanded by 6MB 00:04:11.267 EAL: Calling mem event callback 'spdk:(nil)' 00:04:11.267 EAL: request: mp_malloc_sync 00:04:11.267 EAL: No shared files mode enabled, IPC is disabled 00:04:11.267 EAL: Heap on socket 0 was shrunk by 6MB 00:04:11.267 EAL: Trying to obtain current memory policy. 00:04:11.267 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:11.267 EAL: Restoring previous memory policy: 4 00:04:11.267 EAL: Calling mem event callback 'spdk:(nil)' 00:04:11.267 EAL: request: mp_malloc_sync 00:04:11.267 EAL: No shared files mode enabled, IPC is disabled 00:04:11.267 EAL: Heap on socket 0 was expanded by 10MB 00:04:11.267 EAL: Calling mem event callback 'spdk:(nil)' 00:04:11.267 EAL: request: mp_malloc_sync 00:04:11.267 EAL: No shared files mode enabled, IPC is disabled 00:04:11.267 EAL: Heap on socket 0 was shrunk by 10MB 00:04:11.267 EAL: Trying to obtain current memory policy. 00:04:11.267 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:11.267 EAL: Restoring previous memory policy: 4 00:04:11.267 EAL: Calling mem event callback 'spdk:(nil)' 00:04:11.267 EAL: request: mp_malloc_sync 00:04:11.267 EAL: No shared files mode enabled, IPC is disabled 00:04:11.267 EAL: Heap on socket 0 was expanded by 18MB 00:04:11.267 EAL: Calling mem event callback 'spdk:(nil)' 00:04:11.267 EAL: request: mp_malloc_sync 00:04:11.267 EAL: No shared files mode enabled, IPC is disabled 00:04:11.267 EAL: Heap on socket 0 was shrunk by 18MB 00:04:11.267 EAL: Trying to obtain current memory policy. 00:04:11.267 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:11.267 EAL: Restoring previous memory policy: 4 00:04:11.268 EAL: Calling mem event callback 'spdk:(nil)' 00:04:11.268 EAL: request: mp_malloc_sync 00:04:11.268 EAL: No shared files mode enabled, IPC is disabled 00:04:11.268 EAL: Heap on socket 0 was expanded by 34MB 00:04:11.268 EAL: Calling mem event callback 'spdk:(nil)' 00:04:11.268 EAL: request: mp_malloc_sync 00:04:11.268 EAL: No shared files mode enabled, IPC is disabled 00:04:11.268 EAL: Heap on socket 0 was shrunk by 34MB 00:04:11.268 EAL: Trying to obtain current memory policy. 00:04:11.268 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:11.268 EAL: Restoring previous memory policy: 4 00:04:11.268 EAL: Calling mem event callback 'spdk:(nil)' 00:04:11.268 EAL: request: mp_malloc_sync 00:04:11.268 EAL: No shared files mode enabled, IPC is disabled 00:04:11.268 EAL: Heap on socket 0 was expanded by 66MB 00:04:11.268 EAL: Calling mem event callback 'spdk:(nil)' 00:04:11.268 EAL: request: mp_malloc_sync 00:04:11.268 EAL: No shared files mode enabled, IPC is disabled 00:04:11.268 EAL: Heap on socket 0 was shrunk by 66MB 00:04:11.268 EAL: Trying to obtain current memory policy. 00:04:11.268 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:11.268 EAL: Restoring previous memory policy: 4 00:04:11.268 EAL: Calling mem event callback 'spdk:(nil)' 00:04:11.268 EAL: request: mp_malloc_sync 00:04:11.268 EAL: No shared files mode enabled, IPC is disabled 00:04:11.268 EAL: Heap on socket 0 was expanded by 130MB 00:04:11.268 EAL: Calling mem event callback 'spdk:(nil)' 00:04:11.268 EAL: request: mp_malloc_sync 00:04:11.268 EAL: No shared files mode enabled, IPC is disabled 00:04:11.268 EAL: Heap on socket 0 was shrunk by 130MB 00:04:11.268 EAL: Trying to obtain current memory policy. 00:04:11.268 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:11.268 EAL: Restoring previous memory policy: 4 00:04:11.268 EAL: Calling mem event callback 'spdk:(nil)' 00:04:11.268 EAL: request: mp_malloc_sync 00:04:11.268 EAL: No shared files mode enabled, IPC is disabled 00:04:11.268 EAL: Heap on socket 0 was expanded by 258MB 00:04:11.268 EAL: Calling mem event callback 'spdk:(nil)' 00:04:11.268 EAL: request: mp_malloc_sync 00:04:11.268 EAL: No shared files mode enabled, IPC is disabled 00:04:11.268 EAL: Heap on socket 0 was shrunk by 258MB 00:04:11.268 EAL: Trying to obtain current memory policy. 00:04:11.268 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:11.527 EAL: Restoring previous memory policy: 4 00:04:11.527 EAL: Calling mem event callback 'spdk:(nil)' 00:04:11.527 EAL: request: mp_malloc_sync 00:04:11.527 EAL: No shared files mode enabled, IPC is disabled 00:04:11.527 EAL: Heap on socket 0 was expanded by 514MB 00:04:11.527 EAL: Calling mem event callback 'spdk:(nil)' 00:04:11.527 EAL: request: mp_malloc_sync 00:04:11.527 EAL: No shared files mode enabled, IPC is disabled 00:04:11.527 EAL: Heap on socket 0 was shrunk by 514MB 00:04:11.527 EAL: Trying to obtain current memory policy. 00:04:11.527 EAL: Setting policy MPOL_PREFERRED for socket 0 00:04:11.786 EAL: Restoring previous memory policy: 4 00:04:11.786 EAL: Calling mem event callback 'spdk:(nil)' 00:04:11.786 EAL: request: mp_malloc_sync 00:04:11.786 EAL: No shared files mode enabled, IPC is disabled 00:04:11.786 EAL: Heap on socket 0 was expanded by 1026MB 00:04:12.088 EAL: Calling mem event callback 'spdk:(nil)' 00:04:12.088 EAL: request: mp_malloc_sync 00:04:12.088 EAL: No shared files mode enabled, IPC is disabled 00:04:12.088 EAL: Heap on socket 0 was shrunk by 1026MB 00:04:12.088 passed 00:04:12.088 00:04:12.088 Run Summary: Type Total Ran Passed Failed Inactive 00:04:12.088 suites 1 1 n/a 0 0 00:04:12.088 tests 2 2 2 0 0 00:04:12.088 asserts 497 497 497 0 n/a 00:04:12.088 00:04:12.088 Elapsed time = 0.962 seconds 00:04:12.088 EAL: Calling mem event callback 'spdk:(nil)' 00:04:12.088 EAL: request: mp_malloc_sync 00:04:12.088 EAL: No shared files mode enabled, IPC is disabled 00:04:12.088 EAL: Heap on socket 0 was shrunk by 2MB 00:04:12.088 EAL: No shared files mode enabled, IPC is disabled 00:04:12.088 EAL: No shared files mode enabled, IPC is disabled 00:04:12.088 EAL: No shared files mode enabled, IPC is disabled 00:04:12.088 00:04:12.088 real 0m1.068s 00:04:12.088 user 0m0.631s 00:04:12.088 sys 0m0.412s 00:04:12.088 17:50:05 env.env_vtophys -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:12.088 17:50:05 env.env_vtophys -- common/autotest_common.sh@10 -- # set +x 00:04:12.088 ************************************ 00:04:12.088 END TEST env_vtophys 00:04:12.088 ************************************ 00:04:12.377 17:50:05 env -- common/autotest_common.sh@1142 -- # return 0 00:04:12.377 17:50:05 env -- env/env.sh@12 -- # run_test env_pci /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/pci/pci_ut 00:04:12.377 17:50:05 env -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:12.377 17:50:05 env -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:12.377 17:50:05 env -- common/autotest_common.sh@10 -- # set +x 00:04:12.377 ************************************ 00:04:12.377 START TEST env_pci 00:04:12.377 ************************************ 00:04:12.377 17:50:05 env.env_pci -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/pci/pci_ut 00:04:12.377 00:04:12.377 00:04:12.377 CUnit - A unit testing framework for C - Version 2.1-3 00:04:12.377 http://cunit.sourceforge.net/ 00:04:12.377 00:04:12.377 00:04:12.377 Suite: pci 00:04:12.377 Test: pci_hook ...[2024-07-15 17:50:05.840526] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk/pci.c:1040:spdk_pci_device_claim: *ERROR*: Cannot create lock on device /var/tmp/spdk_pci_lock_10000:00:01.0, probably process 403381 has claimed it 00:04:12.377 EAL: Cannot find device (10000:00:01.0) 00:04:12.377 EAL: Failed to attach device on primary process 00:04:12.377 passed 00:04:12.377 00:04:12.377 Run Summary: Type Total Ran Passed Failed Inactive 00:04:12.377 suites 1 1 n/a 0 0 00:04:12.377 tests 1 1 1 0 0 00:04:12.377 asserts 25 25 25 0 n/a 00:04:12.377 00:04:12.377 Elapsed time = 0.025 seconds 00:04:12.377 00:04:12.377 real 0m0.043s 00:04:12.377 user 0m0.014s 00:04:12.378 sys 0m0.029s 00:04:12.378 17:50:05 env.env_pci -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:12.378 17:50:05 env.env_pci -- common/autotest_common.sh@10 -- # set +x 00:04:12.378 ************************************ 00:04:12.378 END TEST env_pci 00:04:12.378 ************************************ 00:04:12.378 17:50:05 env -- common/autotest_common.sh@1142 -- # return 0 00:04:12.378 17:50:05 env -- env/env.sh@14 -- # argv='-c 0x1 ' 00:04:12.378 17:50:05 env -- env/env.sh@15 -- # uname 00:04:12.378 17:50:05 env -- env/env.sh@15 -- # '[' Linux = Linux ']' 00:04:12.378 17:50:05 env -- env/env.sh@22 -- # argv+=--base-virtaddr=0x200000000000 00:04:12.378 17:50:05 env -- env/env.sh@24 -- # run_test env_dpdk_post_init /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:04:12.378 17:50:05 env -- common/autotest_common.sh@1099 -- # '[' 5 -le 1 ']' 00:04:12.378 17:50:05 env -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:12.378 17:50:05 env -- common/autotest_common.sh@10 -- # set +x 00:04:12.378 ************************************ 00:04:12.378 START TEST env_dpdk_post_init 00:04:12.378 ************************************ 00:04:12.378 17:50:05 env.env_dpdk_post_init -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/env_dpdk_post_init/env_dpdk_post_init -c 0x1 --base-virtaddr=0x200000000000 00:04:12.378 EAL: Detected CPU lcores: 96 00:04:12.378 EAL: Detected NUMA nodes: 2 00:04:12.378 EAL: Detected shared linkage of DPDK 00:04:12.378 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:04:12.378 EAL: Selected IOVA mode 'VA' 00:04:12.378 EAL: No free 2048 kB hugepages reported on node 1 00:04:12.378 EAL: VFIO support initialized 00:04:12.378 TELEMETRY: No legacy callbacks, legacy socket not created 00:04:12.378 EAL: Using IOMMU type 1 (Type 1) 00:04:12.378 EAL: Ignore mapping IO port bar(1) 00:04:12.378 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:00:04.0 (socket 0) 00:04:12.378 EAL: Ignore mapping IO port bar(1) 00:04:12.378 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:00:04.1 (socket 0) 00:04:12.378 EAL: Ignore mapping IO port bar(1) 00:04:12.378 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:00:04.2 (socket 0) 00:04:12.378 EAL: Ignore mapping IO port bar(1) 00:04:12.378 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:00:04.3 (socket 0) 00:04:12.637 EAL: Ignore mapping IO port bar(1) 00:04:12.637 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:00:04.4 (socket 0) 00:04:12.637 EAL: Ignore mapping IO port bar(1) 00:04:12.637 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:00:04.5 (socket 0) 00:04:12.637 EAL: Ignore mapping IO port bar(1) 00:04:12.637 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:00:04.6 (socket 0) 00:04:12.637 EAL: Ignore mapping IO port bar(1) 00:04:12.637 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:00:04.7 (socket 0) 00:04:13.207 EAL: Probe PCI driver: spdk_nvme (8086:0a54) device: 0000:5e:00.0 (socket 0) 00:04:13.207 EAL: Ignore mapping IO port bar(1) 00:04:13.207 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:80:04.0 (socket 1) 00:04:13.207 EAL: Ignore mapping IO port bar(1) 00:04:13.207 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:80:04.1 (socket 1) 00:04:13.207 EAL: Ignore mapping IO port bar(1) 00:04:13.207 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:80:04.2 (socket 1) 00:04:13.207 EAL: Ignore mapping IO port bar(1) 00:04:13.207 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:80:04.3 (socket 1) 00:04:13.466 EAL: Ignore mapping IO port bar(1) 00:04:13.466 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:80:04.4 (socket 1) 00:04:13.466 EAL: Ignore mapping IO port bar(1) 00:04:13.466 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:80:04.5 (socket 1) 00:04:13.466 EAL: Ignore mapping IO port bar(1) 00:04:13.466 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:80:04.6 (socket 1) 00:04:13.466 EAL: Ignore mapping IO port bar(1) 00:04:13.466 EAL: Probe PCI driver: spdk_ioat (8086:2021) device: 0000:80:04.7 (socket 1) 00:04:16.783 EAL: Releasing PCI mapped resource for 0000:5e:00.0 00:04:16.783 EAL: Calling pci_unmap_resource for 0000:5e:00.0 at 0x202001020000 00:04:16.783 Starting DPDK initialization... 00:04:16.783 Starting SPDK post initialization... 00:04:16.783 SPDK NVMe probe 00:04:16.783 Attaching to 0000:5e:00.0 00:04:16.783 Attached to 0000:5e:00.0 00:04:16.783 Cleaning up... 00:04:16.783 00:04:16.783 real 0m4.314s 00:04:16.783 user 0m3.285s 00:04:16.783 sys 0m0.100s 00:04:16.783 17:50:10 env.env_dpdk_post_init -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:16.783 17:50:10 env.env_dpdk_post_init -- common/autotest_common.sh@10 -- # set +x 00:04:16.783 ************************************ 00:04:16.783 END TEST env_dpdk_post_init 00:04:16.783 ************************************ 00:04:16.783 17:50:10 env -- common/autotest_common.sh@1142 -- # return 0 00:04:16.783 17:50:10 env -- env/env.sh@26 -- # uname 00:04:16.783 17:50:10 env -- env/env.sh@26 -- # '[' Linux = Linux ']' 00:04:16.783 17:50:10 env -- env/env.sh@29 -- # run_test env_mem_callbacks /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/mem_callbacks/mem_callbacks 00:04:16.783 17:50:10 env -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:16.783 17:50:10 env -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:16.783 17:50:10 env -- common/autotest_common.sh@10 -- # set +x 00:04:16.783 ************************************ 00:04:16.783 START TEST env_mem_callbacks 00:04:16.783 ************************************ 00:04:16.783 17:50:10 env.env_mem_callbacks -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/env/mem_callbacks/mem_callbacks 00:04:16.783 EAL: Detected CPU lcores: 96 00:04:16.783 EAL: Detected NUMA nodes: 2 00:04:16.783 EAL: Detected shared linkage of DPDK 00:04:16.783 EAL: Multi-process socket /var/run/dpdk/rte/mp_socket 00:04:16.783 EAL: Selected IOVA mode 'VA' 00:04:16.783 EAL: No free 2048 kB hugepages reported on node 1 00:04:16.783 EAL: VFIO support initialized 00:04:16.783 TELEMETRY: No legacy callbacks, legacy socket not created 00:04:16.783 00:04:16.783 00:04:16.783 CUnit - A unit testing framework for C - Version 2.1-3 00:04:16.783 http://cunit.sourceforge.net/ 00:04:16.783 00:04:16.783 00:04:16.783 Suite: memory 00:04:16.783 Test: test ... 00:04:16.783 register 0x200000200000 2097152 00:04:16.783 malloc 3145728 00:04:16.783 register 0x200000400000 4194304 00:04:16.783 buf 0x200000500000 len 3145728 PASSED 00:04:16.783 malloc 64 00:04:16.783 buf 0x2000004fff40 len 64 PASSED 00:04:16.783 malloc 4194304 00:04:16.783 register 0x200000800000 6291456 00:04:16.783 buf 0x200000a00000 len 4194304 PASSED 00:04:16.783 free 0x200000500000 3145728 00:04:16.783 free 0x2000004fff40 64 00:04:16.783 unregister 0x200000400000 4194304 PASSED 00:04:16.783 free 0x200000a00000 4194304 00:04:16.783 unregister 0x200000800000 6291456 PASSED 00:04:16.783 malloc 8388608 00:04:16.783 register 0x200000400000 10485760 00:04:16.783 buf 0x200000600000 len 8388608 PASSED 00:04:16.783 free 0x200000600000 8388608 00:04:16.783 unregister 0x200000400000 10485760 PASSED 00:04:16.783 passed 00:04:16.783 00:04:16.783 Run Summary: Type Total Ran Passed Failed Inactive 00:04:16.783 suites 1 1 n/a 0 0 00:04:16.783 tests 1 1 1 0 0 00:04:16.783 asserts 15 15 15 0 n/a 00:04:16.783 00:04:16.783 Elapsed time = 0.006 seconds 00:04:16.783 00:04:16.783 real 0m0.053s 00:04:16.783 user 0m0.022s 00:04:16.783 sys 0m0.031s 00:04:16.783 17:50:10 env.env_mem_callbacks -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:16.783 17:50:10 env.env_mem_callbacks -- common/autotest_common.sh@10 -- # set +x 00:04:16.783 ************************************ 00:04:16.783 END TEST env_mem_callbacks 00:04:16.783 ************************************ 00:04:16.783 17:50:10 env -- common/autotest_common.sh@1142 -- # return 0 00:04:16.783 00:04:16.783 real 0m6.039s 00:04:16.783 user 0m4.257s 00:04:16.783 sys 0m0.857s 00:04:16.783 17:50:10 env -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:16.783 17:50:10 env -- common/autotest_common.sh@10 -- # set +x 00:04:16.783 ************************************ 00:04:16.783 END TEST env 00:04:16.783 ************************************ 00:04:16.783 17:50:10 -- common/autotest_common.sh@1142 -- # return 0 00:04:16.783 17:50:10 -- spdk/autotest.sh@169 -- # run_test rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/rpc.sh 00:04:16.783 17:50:10 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:16.783 17:50:10 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:16.783 17:50:10 -- common/autotest_common.sh@10 -- # set +x 00:04:16.783 ************************************ 00:04:16.783 START TEST rpc 00:04:16.783 ************************************ 00:04:16.783 17:50:10 rpc -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/rpc.sh 00:04:17.043 * Looking for test storage... 00:04:17.043 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:04:17.043 17:50:10 rpc -- rpc/rpc.sh@65 -- # spdk_pid=404207 00:04:17.043 17:50:10 rpc -- rpc/rpc.sh@66 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:04:17.043 17:50:10 rpc -- rpc/rpc.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -e bdev 00:04:17.043 17:50:10 rpc -- rpc/rpc.sh@67 -- # waitforlisten 404207 00:04:17.043 17:50:10 rpc -- common/autotest_common.sh@829 -- # '[' -z 404207 ']' 00:04:17.043 17:50:10 rpc -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:17.043 17:50:10 rpc -- common/autotest_common.sh@834 -- # local max_retries=100 00:04:17.043 17:50:10 rpc -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:17.043 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:17.043 17:50:10 rpc -- common/autotest_common.sh@838 -- # xtrace_disable 00:04:17.043 17:50:10 rpc -- common/autotest_common.sh@10 -- # set +x 00:04:17.043 [2024-07-15 17:50:10.617577] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:04:17.043 [2024-07-15 17:50:10.617619] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid404207 ] 00:04:17.043 EAL: No free 2048 kB hugepages reported on node 1 00:04:17.043 [2024-07-15 17:50:10.670625] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:17.043 [2024-07-15 17:50:10.744557] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask bdev specified. 00:04:17.043 [2024-07-15 17:50:10.744598] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s spdk_tgt -p 404207' to capture a snapshot of events at runtime. 00:04:17.043 [2024-07-15 17:50:10.744605] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:04:17.044 [2024-07-15 17:50:10.744610] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:04:17.044 [2024-07-15 17:50:10.744615] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/spdk_tgt_trace.pid404207 for offline analysis/debug. 00:04:17.044 [2024-07-15 17:50:10.744650] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:04:17.981 17:50:11 rpc -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:04:17.981 17:50:11 rpc -- common/autotest_common.sh@862 -- # return 0 00:04:17.981 17:50:11 rpc -- rpc/rpc.sh@69 -- # export PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:04:17.981 17:50:11 rpc -- rpc/rpc.sh@69 -- # PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:04:17.981 17:50:11 rpc -- rpc/rpc.sh@72 -- # rpc=rpc_cmd 00:04:17.982 17:50:11 rpc -- rpc/rpc.sh@73 -- # run_test rpc_integrity rpc_integrity 00:04:17.982 17:50:11 rpc -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:17.982 17:50:11 rpc -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:17.982 17:50:11 rpc -- common/autotest_common.sh@10 -- # set +x 00:04:17.982 ************************************ 00:04:17.982 START TEST rpc_integrity 00:04:17.982 ************************************ 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@1123 -- # rpc_integrity 00:04:17.982 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:17.982 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:04:17.982 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # jq length 00:04:17.982 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:04:17.982 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:17.982 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc0 00:04:17.982 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:17.982 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:04:17.982 { 00:04:17.982 "name": "Malloc0", 00:04:17.982 "aliases": [ 00:04:17.982 "a4a58179-8d7c-4dd3-96f6-d14a673d5507" 00:04:17.982 ], 00:04:17.982 "product_name": "Malloc disk", 00:04:17.982 "block_size": 512, 00:04:17.982 "num_blocks": 16384, 00:04:17.982 "uuid": "a4a58179-8d7c-4dd3-96f6-d14a673d5507", 00:04:17.982 "assigned_rate_limits": { 00:04:17.982 "rw_ios_per_sec": 0, 00:04:17.982 "rw_mbytes_per_sec": 0, 00:04:17.982 "r_mbytes_per_sec": 0, 00:04:17.982 "w_mbytes_per_sec": 0 00:04:17.982 }, 00:04:17.982 "claimed": false, 00:04:17.982 "zoned": false, 00:04:17.982 "supported_io_types": { 00:04:17.982 "read": true, 00:04:17.982 "write": true, 00:04:17.982 "unmap": true, 00:04:17.982 "flush": true, 00:04:17.982 "reset": true, 00:04:17.982 "nvme_admin": false, 00:04:17.982 "nvme_io": false, 00:04:17.982 "nvme_io_md": false, 00:04:17.982 "write_zeroes": true, 00:04:17.982 "zcopy": true, 00:04:17.982 "get_zone_info": false, 00:04:17.982 "zone_management": false, 00:04:17.982 "zone_append": false, 00:04:17.982 "compare": false, 00:04:17.982 "compare_and_write": false, 00:04:17.982 "abort": true, 00:04:17.982 "seek_hole": false, 00:04:17.982 "seek_data": false, 00:04:17.982 "copy": true, 00:04:17.982 "nvme_iov_md": false 00:04:17.982 }, 00:04:17.982 "memory_domains": [ 00:04:17.982 { 00:04:17.982 "dma_device_id": "system", 00:04:17.982 "dma_device_type": 1 00:04:17.982 }, 00:04:17.982 { 00:04:17.982 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:17.982 "dma_device_type": 2 00:04:17.982 } 00:04:17.982 ], 00:04:17.982 "driver_specific": {} 00:04:17.982 } 00:04:17.982 ]' 00:04:17.982 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # jq length 00:04:17.982 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:04:17.982 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc0 -p Passthru0 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:17.982 [2024-07-15 17:50:11.570657] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc0 00:04:17.982 [2024-07-15 17:50:11.570688] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:04:17.982 [2024-07-15 17:50:11.570701] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x9ad8c0 00:04:17.982 [2024-07-15 17:50:11.570707] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:04:17.982 [2024-07-15 17:50:11.571799] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:04:17.982 [2024-07-15 17:50:11.571820] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:04:17.982 Passthru0 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:17.982 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:17.982 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:04:17.982 { 00:04:17.982 "name": "Malloc0", 00:04:17.982 "aliases": [ 00:04:17.982 "a4a58179-8d7c-4dd3-96f6-d14a673d5507" 00:04:17.982 ], 00:04:17.982 "product_name": "Malloc disk", 00:04:17.982 "block_size": 512, 00:04:17.982 "num_blocks": 16384, 00:04:17.982 "uuid": "a4a58179-8d7c-4dd3-96f6-d14a673d5507", 00:04:17.982 "assigned_rate_limits": { 00:04:17.982 "rw_ios_per_sec": 0, 00:04:17.982 "rw_mbytes_per_sec": 0, 00:04:17.982 "r_mbytes_per_sec": 0, 00:04:17.982 "w_mbytes_per_sec": 0 00:04:17.982 }, 00:04:17.982 "claimed": true, 00:04:17.982 "claim_type": "exclusive_write", 00:04:17.982 "zoned": false, 00:04:17.982 "supported_io_types": { 00:04:17.982 "read": true, 00:04:17.982 "write": true, 00:04:17.982 "unmap": true, 00:04:17.982 "flush": true, 00:04:17.982 "reset": true, 00:04:17.982 "nvme_admin": false, 00:04:17.982 "nvme_io": false, 00:04:17.982 "nvme_io_md": false, 00:04:17.982 "write_zeroes": true, 00:04:17.982 "zcopy": true, 00:04:17.982 "get_zone_info": false, 00:04:17.982 "zone_management": false, 00:04:17.982 "zone_append": false, 00:04:17.982 "compare": false, 00:04:17.982 "compare_and_write": false, 00:04:17.982 "abort": true, 00:04:17.982 "seek_hole": false, 00:04:17.982 "seek_data": false, 00:04:17.982 "copy": true, 00:04:17.982 "nvme_iov_md": false 00:04:17.982 }, 00:04:17.982 "memory_domains": [ 00:04:17.982 { 00:04:17.982 "dma_device_id": "system", 00:04:17.982 "dma_device_type": 1 00:04:17.982 }, 00:04:17.982 { 00:04:17.982 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:17.982 "dma_device_type": 2 00:04:17.982 } 00:04:17.982 ], 00:04:17.982 "driver_specific": {} 00:04:17.982 }, 00:04:17.982 { 00:04:17.982 "name": "Passthru0", 00:04:17.982 "aliases": [ 00:04:17.982 "2d5a29ec-ed52-576b-95ad-91c4f6ee7f88" 00:04:17.982 ], 00:04:17.982 "product_name": "passthru", 00:04:17.982 "block_size": 512, 00:04:17.982 "num_blocks": 16384, 00:04:17.982 "uuid": "2d5a29ec-ed52-576b-95ad-91c4f6ee7f88", 00:04:17.982 "assigned_rate_limits": { 00:04:17.982 "rw_ios_per_sec": 0, 00:04:17.982 "rw_mbytes_per_sec": 0, 00:04:17.982 "r_mbytes_per_sec": 0, 00:04:17.982 "w_mbytes_per_sec": 0 00:04:17.982 }, 00:04:17.982 "claimed": false, 00:04:17.982 "zoned": false, 00:04:17.982 "supported_io_types": { 00:04:17.982 "read": true, 00:04:17.982 "write": true, 00:04:17.982 "unmap": true, 00:04:17.982 "flush": true, 00:04:17.982 "reset": true, 00:04:17.982 "nvme_admin": false, 00:04:17.982 "nvme_io": false, 00:04:17.982 "nvme_io_md": false, 00:04:17.982 "write_zeroes": true, 00:04:17.982 "zcopy": true, 00:04:17.982 "get_zone_info": false, 00:04:17.982 "zone_management": false, 00:04:17.982 "zone_append": false, 00:04:17.982 "compare": false, 00:04:17.982 "compare_and_write": false, 00:04:17.982 "abort": true, 00:04:17.982 "seek_hole": false, 00:04:17.982 "seek_data": false, 00:04:17.982 "copy": true, 00:04:17.982 "nvme_iov_md": false 00:04:17.982 }, 00:04:17.982 "memory_domains": [ 00:04:17.982 { 00:04:17.982 "dma_device_id": "system", 00:04:17.982 "dma_device_type": 1 00:04:17.982 }, 00:04:17.982 { 00:04:17.982 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:17.982 "dma_device_type": 2 00:04:17.982 } 00:04:17.982 ], 00:04:17.982 "driver_specific": { 00:04:17.982 "passthru": { 00:04:17.982 "name": "Passthru0", 00:04:17.982 "base_bdev_name": "Malloc0" 00:04:17.982 } 00:04:17.982 } 00:04:17.982 } 00:04:17.982 ]' 00:04:17.982 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # jq length 00:04:17.982 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:04:17.982 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:17.982 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc0 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:17.982 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:17.982 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:17.982 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:04:17.982 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # jq length 00:04:18.242 17:50:11 rpc.rpc_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:04:18.242 00:04:18.242 real 0m0.281s 00:04:18.242 user 0m0.185s 00:04:18.242 sys 0m0.029s 00:04:18.242 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:18.242 17:50:11 rpc.rpc_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:18.242 ************************************ 00:04:18.242 END TEST rpc_integrity 00:04:18.242 ************************************ 00:04:18.242 17:50:11 rpc -- common/autotest_common.sh@1142 -- # return 0 00:04:18.242 17:50:11 rpc -- rpc/rpc.sh@74 -- # run_test rpc_plugins rpc_plugins 00:04:18.242 17:50:11 rpc -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:18.242 17:50:11 rpc -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:18.242 17:50:11 rpc -- common/autotest_common.sh@10 -- # set +x 00:04:18.242 ************************************ 00:04:18.242 START TEST rpc_plugins 00:04:18.242 ************************************ 00:04:18.242 17:50:11 rpc.rpc_plugins -- common/autotest_common.sh@1123 -- # rpc_plugins 00:04:18.242 17:50:11 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # rpc_cmd --plugin rpc_plugin create_malloc 00:04:18.242 17:50:11 rpc.rpc_plugins -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:18.242 17:50:11 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:04:18.242 17:50:11 rpc.rpc_plugins -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:18.242 17:50:11 rpc.rpc_plugins -- rpc/rpc.sh@30 -- # malloc=Malloc1 00:04:18.242 17:50:11 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # rpc_cmd bdev_get_bdevs 00:04:18.242 17:50:11 rpc.rpc_plugins -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:18.242 17:50:11 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:04:18.242 17:50:11 rpc.rpc_plugins -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:18.242 17:50:11 rpc.rpc_plugins -- rpc/rpc.sh@31 -- # bdevs='[ 00:04:18.242 { 00:04:18.242 "name": "Malloc1", 00:04:18.242 "aliases": [ 00:04:18.242 "d9b3c0f9-9c1b-48bc-a0ad-f63af32b64f3" 00:04:18.242 ], 00:04:18.242 "product_name": "Malloc disk", 00:04:18.242 "block_size": 4096, 00:04:18.242 "num_blocks": 256, 00:04:18.242 "uuid": "d9b3c0f9-9c1b-48bc-a0ad-f63af32b64f3", 00:04:18.242 "assigned_rate_limits": { 00:04:18.242 "rw_ios_per_sec": 0, 00:04:18.242 "rw_mbytes_per_sec": 0, 00:04:18.242 "r_mbytes_per_sec": 0, 00:04:18.242 "w_mbytes_per_sec": 0 00:04:18.242 }, 00:04:18.242 "claimed": false, 00:04:18.242 "zoned": false, 00:04:18.242 "supported_io_types": { 00:04:18.242 "read": true, 00:04:18.242 "write": true, 00:04:18.242 "unmap": true, 00:04:18.242 "flush": true, 00:04:18.242 "reset": true, 00:04:18.242 "nvme_admin": false, 00:04:18.242 "nvme_io": false, 00:04:18.242 "nvme_io_md": false, 00:04:18.242 "write_zeroes": true, 00:04:18.242 "zcopy": true, 00:04:18.242 "get_zone_info": false, 00:04:18.242 "zone_management": false, 00:04:18.242 "zone_append": false, 00:04:18.242 "compare": false, 00:04:18.242 "compare_and_write": false, 00:04:18.242 "abort": true, 00:04:18.242 "seek_hole": false, 00:04:18.242 "seek_data": false, 00:04:18.242 "copy": true, 00:04:18.242 "nvme_iov_md": false 00:04:18.242 }, 00:04:18.242 "memory_domains": [ 00:04:18.242 { 00:04:18.242 "dma_device_id": "system", 00:04:18.242 "dma_device_type": 1 00:04:18.242 }, 00:04:18.242 { 00:04:18.242 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:18.242 "dma_device_type": 2 00:04:18.242 } 00:04:18.242 ], 00:04:18.242 "driver_specific": {} 00:04:18.242 } 00:04:18.242 ]' 00:04:18.242 17:50:11 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # jq length 00:04:18.242 17:50:11 rpc.rpc_plugins -- rpc/rpc.sh@32 -- # '[' 1 == 1 ']' 00:04:18.242 17:50:11 rpc.rpc_plugins -- rpc/rpc.sh@34 -- # rpc_cmd --plugin rpc_plugin delete_malloc Malloc1 00:04:18.242 17:50:11 rpc.rpc_plugins -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:18.242 17:50:11 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:04:18.242 17:50:11 rpc.rpc_plugins -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:18.242 17:50:11 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # rpc_cmd bdev_get_bdevs 00:04:18.242 17:50:11 rpc.rpc_plugins -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:18.242 17:50:11 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:04:18.242 17:50:11 rpc.rpc_plugins -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:18.242 17:50:11 rpc.rpc_plugins -- rpc/rpc.sh@35 -- # bdevs='[]' 00:04:18.242 17:50:11 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # jq length 00:04:18.242 17:50:11 rpc.rpc_plugins -- rpc/rpc.sh@36 -- # '[' 0 == 0 ']' 00:04:18.242 00:04:18.242 real 0m0.128s 00:04:18.242 user 0m0.080s 00:04:18.242 sys 0m0.013s 00:04:18.242 17:50:11 rpc.rpc_plugins -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:18.242 17:50:11 rpc.rpc_plugins -- common/autotest_common.sh@10 -- # set +x 00:04:18.242 ************************************ 00:04:18.242 END TEST rpc_plugins 00:04:18.242 ************************************ 00:04:18.242 17:50:11 rpc -- common/autotest_common.sh@1142 -- # return 0 00:04:18.242 17:50:11 rpc -- rpc/rpc.sh@75 -- # run_test rpc_trace_cmd_test rpc_trace_cmd_test 00:04:18.243 17:50:11 rpc -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:18.243 17:50:11 rpc -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:18.243 17:50:11 rpc -- common/autotest_common.sh@10 -- # set +x 00:04:18.502 ************************************ 00:04:18.502 START TEST rpc_trace_cmd_test 00:04:18.502 ************************************ 00:04:18.502 17:50:11 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1123 -- # rpc_trace_cmd_test 00:04:18.502 17:50:11 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@40 -- # local info 00:04:18.502 17:50:11 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # rpc_cmd trace_get_info 00:04:18.502 17:50:11 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:18.502 17:50:11 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:04:18.502 17:50:11 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:18.502 17:50:11 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@42 -- # info='{ 00:04:18.502 "tpoint_shm_path": "/dev/shm/spdk_tgt_trace.pid404207", 00:04:18.502 "tpoint_group_mask": "0x8", 00:04:18.502 "iscsi_conn": { 00:04:18.502 "mask": "0x2", 00:04:18.502 "tpoint_mask": "0x0" 00:04:18.502 }, 00:04:18.502 "scsi": { 00:04:18.502 "mask": "0x4", 00:04:18.502 "tpoint_mask": "0x0" 00:04:18.502 }, 00:04:18.502 "bdev": { 00:04:18.502 "mask": "0x8", 00:04:18.502 "tpoint_mask": "0xffffffffffffffff" 00:04:18.502 }, 00:04:18.502 "nvmf_rdma": { 00:04:18.502 "mask": "0x10", 00:04:18.502 "tpoint_mask": "0x0" 00:04:18.502 }, 00:04:18.502 "nvmf_tcp": { 00:04:18.502 "mask": "0x20", 00:04:18.502 "tpoint_mask": "0x0" 00:04:18.502 }, 00:04:18.502 "ftl": { 00:04:18.502 "mask": "0x40", 00:04:18.502 "tpoint_mask": "0x0" 00:04:18.502 }, 00:04:18.502 "blobfs": { 00:04:18.502 "mask": "0x80", 00:04:18.502 "tpoint_mask": "0x0" 00:04:18.502 }, 00:04:18.502 "dsa": { 00:04:18.502 "mask": "0x200", 00:04:18.502 "tpoint_mask": "0x0" 00:04:18.502 }, 00:04:18.502 "thread": { 00:04:18.502 "mask": "0x400", 00:04:18.502 "tpoint_mask": "0x0" 00:04:18.502 }, 00:04:18.502 "nvme_pcie": { 00:04:18.502 "mask": "0x800", 00:04:18.502 "tpoint_mask": "0x0" 00:04:18.502 }, 00:04:18.502 "iaa": { 00:04:18.502 "mask": "0x1000", 00:04:18.502 "tpoint_mask": "0x0" 00:04:18.502 }, 00:04:18.502 "nvme_tcp": { 00:04:18.502 "mask": "0x2000", 00:04:18.502 "tpoint_mask": "0x0" 00:04:18.502 }, 00:04:18.502 "bdev_nvme": { 00:04:18.502 "mask": "0x4000", 00:04:18.502 "tpoint_mask": "0x0" 00:04:18.502 }, 00:04:18.502 "sock": { 00:04:18.502 "mask": "0x8000", 00:04:18.502 "tpoint_mask": "0x0" 00:04:18.502 } 00:04:18.502 }' 00:04:18.502 17:50:11 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # jq length 00:04:18.502 17:50:12 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@43 -- # '[' 16 -gt 2 ']' 00:04:18.502 17:50:12 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # jq 'has("tpoint_group_mask")' 00:04:18.502 17:50:12 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@44 -- # '[' true = true ']' 00:04:18.502 17:50:12 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # jq 'has("tpoint_shm_path")' 00:04:18.502 17:50:12 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@45 -- # '[' true = true ']' 00:04:18.502 17:50:12 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # jq 'has("bdev")' 00:04:18.502 17:50:12 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@46 -- # '[' true = true ']' 00:04:18.502 17:50:12 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # jq -r .bdev.tpoint_mask 00:04:18.502 17:50:12 rpc.rpc_trace_cmd_test -- rpc/rpc.sh@47 -- # '[' 0xffffffffffffffff '!=' 0x0 ']' 00:04:18.502 00:04:18.502 real 0m0.214s 00:04:18.502 user 0m0.182s 00:04:18.502 sys 0m0.024s 00:04:18.502 17:50:12 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:18.502 17:50:12 rpc.rpc_trace_cmd_test -- common/autotest_common.sh@10 -- # set +x 00:04:18.502 ************************************ 00:04:18.502 END TEST rpc_trace_cmd_test 00:04:18.502 ************************************ 00:04:18.502 17:50:12 rpc -- common/autotest_common.sh@1142 -- # return 0 00:04:18.502 17:50:12 rpc -- rpc/rpc.sh@76 -- # [[ 0 -eq 1 ]] 00:04:18.502 17:50:12 rpc -- rpc/rpc.sh@80 -- # rpc=rpc_cmd 00:04:18.502 17:50:12 rpc -- rpc/rpc.sh@81 -- # run_test rpc_daemon_integrity rpc_integrity 00:04:18.502 17:50:12 rpc -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:18.502 17:50:12 rpc -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:18.502 17:50:12 rpc -- common/autotest_common.sh@10 -- # set +x 00:04:18.762 ************************************ 00:04:18.762 START TEST rpc_daemon_integrity 00:04:18.762 ************************************ 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1123 -- # rpc_integrity 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # rpc_cmd bdev_get_bdevs 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@12 -- # bdevs='[]' 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # jq length 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@13 -- # '[' 0 == 0 ']' 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # rpc_cmd bdev_malloc_create 8 512 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@15 -- # malloc=Malloc2 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # rpc_cmd bdev_get_bdevs 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@16 -- # bdevs='[ 00:04:18.762 { 00:04:18.762 "name": "Malloc2", 00:04:18.762 "aliases": [ 00:04:18.762 "114754cc-8b8f-4c1c-ac08-8852839e47c4" 00:04:18.762 ], 00:04:18.762 "product_name": "Malloc disk", 00:04:18.762 "block_size": 512, 00:04:18.762 "num_blocks": 16384, 00:04:18.762 "uuid": "114754cc-8b8f-4c1c-ac08-8852839e47c4", 00:04:18.762 "assigned_rate_limits": { 00:04:18.762 "rw_ios_per_sec": 0, 00:04:18.762 "rw_mbytes_per_sec": 0, 00:04:18.762 "r_mbytes_per_sec": 0, 00:04:18.762 "w_mbytes_per_sec": 0 00:04:18.762 }, 00:04:18.762 "claimed": false, 00:04:18.762 "zoned": false, 00:04:18.762 "supported_io_types": { 00:04:18.762 "read": true, 00:04:18.762 "write": true, 00:04:18.762 "unmap": true, 00:04:18.762 "flush": true, 00:04:18.762 "reset": true, 00:04:18.762 "nvme_admin": false, 00:04:18.762 "nvme_io": false, 00:04:18.762 "nvme_io_md": false, 00:04:18.762 "write_zeroes": true, 00:04:18.762 "zcopy": true, 00:04:18.762 "get_zone_info": false, 00:04:18.762 "zone_management": false, 00:04:18.762 "zone_append": false, 00:04:18.762 "compare": false, 00:04:18.762 "compare_and_write": false, 00:04:18.762 "abort": true, 00:04:18.762 "seek_hole": false, 00:04:18.762 "seek_data": false, 00:04:18.762 "copy": true, 00:04:18.762 "nvme_iov_md": false 00:04:18.762 }, 00:04:18.762 "memory_domains": [ 00:04:18.762 { 00:04:18.762 "dma_device_id": "system", 00:04:18.762 "dma_device_type": 1 00:04:18.762 }, 00:04:18.762 { 00:04:18.762 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:18.762 "dma_device_type": 2 00:04:18.762 } 00:04:18.762 ], 00:04:18.762 "driver_specific": {} 00:04:18.762 } 00:04:18.762 ]' 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # jq length 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@17 -- # '[' 1 == 1 ']' 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@19 -- # rpc_cmd bdev_passthru_create -b Malloc2 -p Passthru0 00:04:18.762 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:18.763 [2024-07-15 17:50:12.376856] vbdev_passthru.c: 607:vbdev_passthru_register: *NOTICE*: Match on Malloc2 00:04:18.763 [2024-07-15 17:50:12.376885] vbdev_passthru.c: 635:vbdev_passthru_register: *NOTICE*: base bdev opened 00:04:18.763 [2024-07-15 17:50:12.376900] vbdev_passthru.c: 680:vbdev_passthru_register: *NOTICE*: io_device created at: 0x0x7fbfb0 00:04:18.763 [2024-07-15 17:50:12.376907] vbdev_passthru.c: 695:vbdev_passthru_register: *NOTICE*: bdev claimed 00:04:18.763 [2024-07-15 17:50:12.377866] vbdev_passthru.c: 708:vbdev_passthru_register: *NOTICE*: pt_bdev registered 00:04:18.763 [2024-07-15 17:50:12.377886] vbdev_passthru.c: 709:vbdev_passthru_register: *NOTICE*: created pt_bdev for: Passthru0 00:04:18.763 Passthru0 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # rpc_cmd bdev_get_bdevs 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@20 -- # bdevs='[ 00:04:18.763 { 00:04:18.763 "name": "Malloc2", 00:04:18.763 "aliases": [ 00:04:18.763 "114754cc-8b8f-4c1c-ac08-8852839e47c4" 00:04:18.763 ], 00:04:18.763 "product_name": "Malloc disk", 00:04:18.763 "block_size": 512, 00:04:18.763 "num_blocks": 16384, 00:04:18.763 "uuid": "114754cc-8b8f-4c1c-ac08-8852839e47c4", 00:04:18.763 "assigned_rate_limits": { 00:04:18.763 "rw_ios_per_sec": 0, 00:04:18.763 "rw_mbytes_per_sec": 0, 00:04:18.763 "r_mbytes_per_sec": 0, 00:04:18.763 "w_mbytes_per_sec": 0 00:04:18.763 }, 00:04:18.763 "claimed": true, 00:04:18.763 "claim_type": "exclusive_write", 00:04:18.763 "zoned": false, 00:04:18.763 "supported_io_types": { 00:04:18.763 "read": true, 00:04:18.763 "write": true, 00:04:18.763 "unmap": true, 00:04:18.763 "flush": true, 00:04:18.763 "reset": true, 00:04:18.763 "nvme_admin": false, 00:04:18.763 "nvme_io": false, 00:04:18.763 "nvme_io_md": false, 00:04:18.763 "write_zeroes": true, 00:04:18.763 "zcopy": true, 00:04:18.763 "get_zone_info": false, 00:04:18.763 "zone_management": false, 00:04:18.763 "zone_append": false, 00:04:18.763 "compare": false, 00:04:18.763 "compare_and_write": false, 00:04:18.763 "abort": true, 00:04:18.763 "seek_hole": false, 00:04:18.763 "seek_data": false, 00:04:18.763 "copy": true, 00:04:18.763 "nvme_iov_md": false 00:04:18.763 }, 00:04:18.763 "memory_domains": [ 00:04:18.763 { 00:04:18.763 "dma_device_id": "system", 00:04:18.763 "dma_device_type": 1 00:04:18.763 }, 00:04:18.763 { 00:04:18.763 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:18.763 "dma_device_type": 2 00:04:18.763 } 00:04:18.763 ], 00:04:18.763 "driver_specific": {} 00:04:18.763 }, 00:04:18.763 { 00:04:18.763 "name": "Passthru0", 00:04:18.763 "aliases": [ 00:04:18.763 "5f85d41f-b41b-5ad2-b6e8-acb98438e5c1" 00:04:18.763 ], 00:04:18.763 "product_name": "passthru", 00:04:18.763 "block_size": 512, 00:04:18.763 "num_blocks": 16384, 00:04:18.763 "uuid": "5f85d41f-b41b-5ad2-b6e8-acb98438e5c1", 00:04:18.763 "assigned_rate_limits": { 00:04:18.763 "rw_ios_per_sec": 0, 00:04:18.763 "rw_mbytes_per_sec": 0, 00:04:18.763 "r_mbytes_per_sec": 0, 00:04:18.763 "w_mbytes_per_sec": 0 00:04:18.763 }, 00:04:18.763 "claimed": false, 00:04:18.763 "zoned": false, 00:04:18.763 "supported_io_types": { 00:04:18.763 "read": true, 00:04:18.763 "write": true, 00:04:18.763 "unmap": true, 00:04:18.763 "flush": true, 00:04:18.763 "reset": true, 00:04:18.763 "nvme_admin": false, 00:04:18.763 "nvme_io": false, 00:04:18.763 "nvme_io_md": false, 00:04:18.763 "write_zeroes": true, 00:04:18.763 "zcopy": true, 00:04:18.763 "get_zone_info": false, 00:04:18.763 "zone_management": false, 00:04:18.763 "zone_append": false, 00:04:18.763 "compare": false, 00:04:18.763 "compare_and_write": false, 00:04:18.763 "abort": true, 00:04:18.763 "seek_hole": false, 00:04:18.763 "seek_data": false, 00:04:18.763 "copy": true, 00:04:18.763 "nvme_iov_md": false 00:04:18.763 }, 00:04:18.763 "memory_domains": [ 00:04:18.763 { 00:04:18.763 "dma_device_id": "system", 00:04:18.763 "dma_device_type": 1 00:04:18.763 }, 00:04:18.763 { 00:04:18.763 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:04:18.763 "dma_device_type": 2 00:04:18.763 } 00:04:18.763 ], 00:04:18.763 "driver_specific": { 00:04:18.763 "passthru": { 00:04:18.763 "name": "Passthru0", 00:04:18.763 "base_bdev_name": "Malloc2" 00:04:18.763 } 00:04:18.763 } 00:04:18.763 } 00:04:18.763 ]' 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # jq length 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@21 -- # '[' 2 == 2 ']' 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@23 -- # rpc_cmd bdev_passthru_delete Passthru0 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@24 -- # rpc_cmd bdev_malloc_delete Malloc2 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # rpc_cmd bdev_get_bdevs 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@25 -- # bdevs='[]' 00:04:18.763 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # jq length 00:04:19.022 17:50:12 rpc.rpc_daemon_integrity -- rpc/rpc.sh@26 -- # '[' 0 == 0 ']' 00:04:19.023 00:04:19.023 real 0m0.270s 00:04:19.023 user 0m0.168s 00:04:19.023 sys 0m0.038s 00:04:19.023 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:19.023 17:50:12 rpc.rpc_daemon_integrity -- common/autotest_common.sh@10 -- # set +x 00:04:19.023 ************************************ 00:04:19.023 END TEST rpc_daemon_integrity 00:04:19.023 ************************************ 00:04:19.023 17:50:12 rpc -- common/autotest_common.sh@1142 -- # return 0 00:04:19.023 17:50:12 rpc -- rpc/rpc.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:04:19.023 17:50:12 rpc -- rpc/rpc.sh@84 -- # killprocess 404207 00:04:19.023 17:50:12 rpc -- common/autotest_common.sh@948 -- # '[' -z 404207 ']' 00:04:19.023 17:50:12 rpc -- common/autotest_common.sh@952 -- # kill -0 404207 00:04:19.023 17:50:12 rpc -- common/autotest_common.sh@953 -- # uname 00:04:19.023 17:50:12 rpc -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:04:19.023 17:50:12 rpc -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 404207 00:04:19.023 17:50:12 rpc -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:04:19.023 17:50:12 rpc -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:04:19.023 17:50:12 rpc -- common/autotest_common.sh@966 -- # echo 'killing process with pid 404207' 00:04:19.023 killing process with pid 404207 00:04:19.023 17:50:12 rpc -- common/autotest_common.sh@967 -- # kill 404207 00:04:19.023 17:50:12 rpc -- common/autotest_common.sh@972 -- # wait 404207 00:04:19.282 00:04:19.282 real 0m2.427s 00:04:19.282 user 0m3.131s 00:04:19.282 sys 0m0.646s 00:04:19.282 17:50:12 rpc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:19.282 17:50:12 rpc -- common/autotest_common.sh@10 -- # set +x 00:04:19.282 ************************************ 00:04:19.282 END TEST rpc 00:04:19.282 ************************************ 00:04:19.282 17:50:12 -- common/autotest_common.sh@1142 -- # return 0 00:04:19.282 17:50:12 -- spdk/autotest.sh@170 -- # run_test skip_rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/skip_rpc.sh 00:04:19.282 17:50:12 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:19.282 17:50:12 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:19.282 17:50:12 -- common/autotest_common.sh@10 -- # set +x 00:04:19.282 ************************************ 00:04:19.282 START TEST skip_rpc 00:04:19.282 ************************************ 00:04:19.282 17:50:12 skip_rpc -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/skip_rpc.sh 00:04:19.542 * Looking for test storage... 00:04:19.542 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc 00:04:19.542 17:50:13 skip_rpc -- rpc/skip_rpc.sh@11 -- # CONFIG_PATH=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:04:19.542 17:50:13 skip_rpc -- rpc/skip_rpc.sh@12 -- # LOG_PATH=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/log.txt 00:04:19.542 17:50:13 skip_rpc -- rpc/skip_rpc.sh@73 -- # run_test skip_rpc test_skip_rpc 00:04:19.542 17:50:13 skip_rpc -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:19.542 17:50:13 skip_rpc -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:19.542 17:50:13 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:19.542 ************************************ 00:04:19.542 START TEST skip_rpc 00:04:19.542 ************************************ 00:04:19.542 17:50:13 skip_rpc.skip_rpc -- common/autotest_common.sh@1123 -- # test_skip_rpc 00:04:19.542 17:50:13 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@16 -- # local spdk_pid=404836 00:04:19.542 17:50:13 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@18 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:04:19.542 17:50:13 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 00:04:19.542 17:50:13 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@19 -- # sleep 5 00:04:19.542 [2024-07-15 17:50:13.150158] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:04:19.542 [2024-07-15 17:50:13.150195] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid404836 ] 00:04:19.542 EAL: No free 2048 kB hugepages reported on node 1 00:04:19.542 [2024-07-15 17:50:13.201158] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:19.800 [2024-07-15 17:50:13.273475] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@21 -- # NOT rpc_cmd spdk_get_version 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@648 -- # local es=0 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@650 -- # valid_exec_arg rpc_cmd spdk_get_version 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@636 -- # local arg=rpc_cmd 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@640 -- # type -t rpc_cmd 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@651 -- # rpc_cmd spdk_get_version 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@587 -- # [[ 1 == 0 ]] 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@651 -- # es=1 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@22 -- # trap - SIGINT SIGTERM EXIT 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- rpc/skip_rpc.sh@23 -- # killprocess 404836 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@948 -- # '[' -z 404836 ']' 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@952 -- # kill -0 404836 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@953 -- # uname 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 404836 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@966 -- # echo 'killing process with pid 404836' 00:04:25.075 killing process with pid 404836 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@967 -- # kill 404836 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@972 -- # wait 404836 00:04:25.075 00:04:25.075 real 0m5.360s 00:04:25.075 user 0m5.140s 00:04:25.075 sys 0m0.247s 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:25.075 17:50:18 skip_rpc.skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:25.075 ************************************ 00:04:25.075 END TEST skip_rpc 00:04:25.075 ************************************ 00:04:25.075 17:50:18 skip_rpc -- common/autotest_common.sh@1142 -- # return 0 00:04:25.075 17:50:18 skip_rpc -- rpc/skip_rpc.sh@74 -- # run_test skip_rpc_with_json test_skip_rpc_with_json 00:04:25.075 17:50:18 skip_rpc -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:25.075 17:50:18 skip_rpc -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:25.075 17:50:18 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:25.075 ************************************ 00:04:25.075 START TEST skip_rpc_with_json 00:04:25.075 ************************************ 00:04:25.075 17:50:18 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1123 -- # test_skip_rpc_with_json 00:04:25.075 17:50:18 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@44 -- # gen_json_config 00:04:25.075 17:50:18 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:04:25.075 17:50:18 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@28 -- # local spdk_pid=405783 00:04:25.075 17:50:18 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@30 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:04:25.075 17:50:18 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@31 -- # waitforlisten 405783 00:04:25.075 17:50:18 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@829 -- # '[' -z 405783 ']' 00:04:25.075 17:50:18 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:25.075 17:50:18 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@834 -- # local max_retries=100 00:04:25.075 17:50:18 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:25.075 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:25.075 17:50:18 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@838 -- # xtrace_disable 00:04:25.075 17:50:18 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:04:25.075 [2024-07-15 17:50:18.562476] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:04:25.075 [2024-07-15 17:50:18.562513] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid405783 ] 00:04:25.075 EAL: No free 2048 kB hugepages reported on node 1 00:04:25.075 [2024-07-15 17:50:18.614798] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:25.075 [2024-07-15 17:50:18.692708] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:04:25.358 17:50:18 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:04:25.358 17:50:18 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@862 -- # return 0 00:04:25.358 17:50:18 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_get_transports --trtype tcp 00:04:25.358 17:50:18 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:25.358 17:50:18 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:04:25.358 [2024-07-15 17:50:18.883716] nvmf_rpc.c:2562:rpc_nvmf_get_transports: *ERROR*: transport 'tcp' does not exist 00:04:25.358 request: 00:04:25.358 { 00:04:25.358 "trtype": "tcp", 00:04:25.358 "method": "nvmf_get_transports", 00:04:25.358 "req_id": 1 00:04:25.358 } 00:04:25.358 Got JSON-RPC error response 00:04:25.358 response: 00:04:25.358 { 00:04:25.358 "code": -19, 00:04:25.358 "message": "No such device" 00:04:25.358 } 00:04:25.358 17:50:18 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@587 -- # [[ 1 == 0 ]] 00:04:25.358 17:50:18 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@34 -- # rpc_cmd nvmf_create_transport -t tcp 00:04:25.358 17:50:18 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:25.358 17:50:18 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:04:25.358 [2024-07-15 17:50:18.895824] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:04:25.358 17:50:18 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:25.358 17:50:18 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@36 -- # rpc_cmd save_config 00:04:25.358 17:50:18 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:25.358 17:50:18 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:04:25.358 17:50:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:25.358 17:50:19 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@37 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:04:25.358 { 00:04:25.358 "subsystems": [ 00:04:25.358 { 00:04:25.358 "subsystem": "vfio_user_target", 00:04:25.358 "config": null 00:04:25.358 }, 00:04:25.358 { 00:04:25.358 "subsystem": "keyring", 00:04:25.358 "config": [] 00:04:25.358 }, 00:04:25.358 { 00:04:25.358 "subsystem": "iobuf", 00:04:25.358 "config": [ 00:04:25.358 { 00:04:25.358 "method": "iobuf_set_options", 00:04:25.358 "params": { 00:04:25.358 "small_pool_count": 8192, 00:04:25.358 "large_pool_count": 1024, 00:04:25.358 "small_bufsize": 8192, 00:04:25.358 "large_bufsize": 135168 00:04:25.358 } 00:04:25.358 } 00:04:25.358 ] 00:04:25.358 }, 00:04:25.358 { 00:04:25.358 "subsystem": "sock", 00:04:25.358 "config": [ 00:04:25.358 { 00:04:25.358 "method": "sock_set_default_impl", 00:04:25.358 "params": { 00:04:25.358 "impl_name": "posix" 00:04:25.358 } 00:04:25.358 }, 00:04:25.358 { 00:04:25.358 "method": "sock_impl_set_options", 00:04:25.358 "params": { 00:04:25.358 "impl_name": "ssl", 00:04:25.358 "recv_buf_size": 4096, 00:04:25.358 "send_buf_size": 4096, 00:04:25.358 "enable_recv_pipe": true, 00:04:25.358 "enable_quickack": false, 00:04:25.358 "enable_placement_id": 0, 00:04:25.358 "enable_zerocopy_send_server": true, 00:04:25.358 "enable_zerocopy_send_client": false, 00:04:25.358 "zerocopy_threshold": 0, 00:04:25.358 "tls_version": 0, 00:04:25.358 "enable_ktls": false 00:04:25.358 } 00:04:25.358 }, 00:04:25.358 { 00:04:25.358 "method": "sock_impl_set_options", 00:04:25.358 "params": { 00:04:25.358 "impl_name": "posix", 00:04:25.358 "recv_buf_size": 2097152, 00:04:25.358 "send_buf_size": 2097152, 00:04:25.358 "enable_recv_pipe": true, 00:04:25.358 "enable_quickack": false, 00:04:25.358 "enable_placement_id": 0, 00:04:25.358 "enable_zerocopy_send_server": true, 00:04:25.359 "enable_zerocopy_send_client": false, 00:04:25.359 "zerocopy_threshold": 0, 00:04:25.359 "tls_version": 0, 00:04:25.359 "enable_ktls": false 00:04:25.359 } 00:04:25.359 } 00:04:25.359 ] 00:04:25.359 }, 00:04:25.359 { 00:04:25.359 "subsystem": "vmd", 00:04:25.359 "config": [] 00:04:25.359 }, 00:04:25.359 { 00:04:25.359 "subsystem": "accel", 00:04:25.359 "config": [ 00:04:25.359 { 00:04:25.359 "method": "accel_set_options", 00:04:25.359 "params": { 00:04:25.359 "small_cache_size": 128, 00:04:25.359 "large_cache_size": 16, 00:04:25.359 "task_count": 2048, 00:04:25.359 "sequence_count": 2048, 00:04:25.359 "buf_count": 2048 00:04:25.359 } 00:04:25.359 } 00:04:25.359 ] 00:04:25.359 }, 00:04:25.359 { 00:04:25.359 "subsystem": "bdev", 00:04:25.359 "config": [ 00:04:25.359 { 00:04:25.359 "method": "bdev_set_options", 00:04:25.359 "params": { 00:04:25.359 "bdev_io_pool_size": 65535, 00:04:25.359 "bdev_io_cache_size": 256, 00:04:25.359 "bdev_auto_examine": true, 00:04:25.359 "iobuf_small_cache_size": 128, 00:04:25.359 "iobuf_large_cache_size": 16 00:04:25.359 } 00:04:25.359 }, 00:04:25.359 { 00:04:25.359 "method": "bdev_raid_set_options", 00:04:25.359 "params": { 00:04:25.359 "process_window_size_kb": 1024 00:04:25.359 } 00:04:25.359 }, 00:04:25.359 { 00:04:25.359 "method": "bdev_iscsi_set_options", 00:04:25.359 "params": { 00:04:25.359 "timeout_sec": 30 00:04:25.359 } 00:04:25.359 }, 00:04:25.359 { 00:04:25.359 "method": "bdev_nvme_set_options", 00:04:25.359 "params": { 00:04:25.359 "action_on_timeout": "none", 00:04:25.359 "timeout_us": 0, 00:04:25.359 "timeout_admin_us": 0, 00:04:25.359 "keep_alive_timeout_ms": 10000, 00:04:25.359 "arbitration_burst": 0, 00:04:25.359 "low_priority_weight": 0, 00:04:25.359 "medium_priority_weight": 0, 00:04:25.359 "high_priority_weight": 0, 00:04:25.359 "nvme_adminq_poll_period_us": 10000, 00:04:25.359 "nvme_ioq_poll_period_us": 0, 00:04:25.359 "io_queue_requests": 0, 00:04:25.359 "delay_cmd_submit": true, 00:04:25.359 "transport_retry_count": 4, 00:04:25.359 "bdev_retry_count": 3, 00:04:25.359 "transport_ack_timeout": 0, 00:04:25.359 "ctrlr_loss_timeout_sec": 0, 00:04:25.359 "reconnect_delay_sec": 0, 00:04:25.359 "fast_io_fail_timeout_sec": 0, 00:04:25.359 "disable_auto_failback": false, 00:04:25.359 "generate_uuids": false, 00:04:25.359 "transport_tos": 0, 00:04:25.359 "nvme_error_stat": false, 00:04:25.359 "rdma_srq_size": 0, 00:04:25.359 "io_path_stat": false, 00:04:25.359 "allow_accel_sequence": false, 00:04:25.359 "rdma_max_cq_size": 0, 00:04:25.359 "rdma_cm_event_timeout_ms": 0, 00:04:25.359 "dhchap_digests": [ 00:04:25.359 "sha256", 00:04:25.359 "sha384", 00:04:25.359 "sha512" 00:04:25.359 ], 00:04:25.359 "dhchap_dhgroups": [ 00:04:25.359 "null", 00:04:25.359 "ffdhe2048", 00:04:25.359 "ffdhe3072", 00:04:25.359 "ffdhe4096", 00:04:25.359 "ffdhe6144", 00:04:25.359 "ffdhe8192" 00:04:25.359 ] 00:04:25.359 } 00:04:25.359 }, 00:04:25.359 { 00:04:25.359 "method": "bdev_nvme_set_hotplug", 00:04:25.359 "params": { 00:04:25.359 "period_us": 100000, 00:04:25.359 "enable": false 00:04:25.359 } 00:04:25.359 }, 00:04:25.359 { 00:04:25.359 "method": "bdev_wait_for_examine" 00:04:25.359 } 00:04:25.359 ] 00:04:25.359 }, 00:04:25.359 { 00:04:25.359 "subsystem": "scsi", 00:04:25.359 "config": null 00:04:25.359 }, 00:04:25.359 { 00:04:25.359 "subsystem": "scheduler", 00:04:25.359 "config": [ 00:04:25.359 { 00:04:25.359 "method": "framework_set_scheduler", 00:04:25.359 "params": { 00:04:25.359 "name": "static" 00:04:25.359 } 00:04:25.359 } 00:04:25.359 ] 00:04:25.359 }, 00:04:25.359 { 00:04:25.359 "subsystem": "vhost_scsi", 00:04:25.359 "config": [] 00:04:25.359 }, 00:04:25.359 { 00:04:25.359 "subsystem": "vhost_blk", 00:04:25.359 "config": [] 00:04:25.359 }, 00:04:25.359 { 00:04:25.359 "subsystem": "ublk", 00:04:25.359 "config": [] 00:04:25.359 }, 00:04:25.359 { 00:04:25.359 "subsystem": "nbd", 00:04:25.359 "config": [] 00:04:25.359 }, 00:04:25.359 { 00:04:25.359 "subsystem": "nvmf", 00:04:25.359 "config": [ 00:04:25.359 { 00:04:25.359 "method": "nvmf_set_config", 00:04:25.359 "params": { 00:04:25.359 "discovery_filter": "match_any", 00:04:25.359 "admin_cmd_passthru": { 00:04:25.359 "identify_ctrlr": false 00:04:25.359 } 00:04:25.359 } 00:04:25.359 }, 00:04:25.359 { 00:04:25.359 "method": "nvmf_set_max_subsystems", 00:04:25.359 "params": { 00:04:25.359 "max_subsystems": 1024 00:04:25.359 } 00:04:25.359 }, 00:04:25.359 { 00:04:25.359 "method": "nvmf_set_crdt", 00:04:25.359 "params": { 00:04:25.359 "crdt1": 0, 00:04:25.359 "crdt2": 0, 00:04:25.359 "crdt3": 0 00:04:25.359 } 00:04:25.359 }, 00:04:25.359 { 00:04:25.359 "method": "nvmf_create_transport", 00:04:25.359 "params": { 00:04:25.359 "trtype": "TCP", 00:04:25.359 "max_queue_depth": 128, 00:04:25.359 "max_io_qpairs_per_ctrlr": 127, 00:04:25.359 "in_capsule_data_size": 4096, 00:04:25.359 "max_io_size": 131072, 00:04:25.359 "io_unit_size": 131072, 00:04:25.359 "max_aq_depth": 128, 00:04:25.359 "num_shared_buffers": 511, 00:04:25.359 "buf_cache_size": 4294967295, 00:04:25.359 "dif_insert_or_strip": false, 00:04:25.359 "zcopy": false, 00:04:25.359 "c2h_success": true, 00:04:25.359 "sock_priority": 0, 00:04:25.359 "abort_timeout_sec": 1, 00:04:25.359 "ack_timeout": 0, 00:04:25.359 "data_wr_pool_size": 0 00:04:25.359 } 00:04:25.359 } 00:04:25.359 ] 00:04:25.359 }, 00:04:25.359 { 00:04:25.359 "subsystem": "iscsi", 00:04:25.359 "config": [ 00:04:25.359 { 00:04:25.359 "method": "iscsi_set_options", 00:04:25.359 "params": { 00:04:25.359 "node_base": "iqn.2016-06.io.spdk", 00:04:25.359 "max_sessions": 128, 00:04:25.359 "max_connections_per_session": 2, 00:04:25.359 "max_queue_depth": 64, 00:04:25.359 "default_time2wait": 2, 00:04:25.359 "default_time2retain": 20, 00:04:25.359 "first_burst_length": 8192, 00:04:25.359 "immediate_data": true, 00:04:25.359 "allow_duplicated_isid": false, 00:04:25.359 "error_recovery_level": 0, 00:04:25.359 "nop_timeout": 60, 00:04:25.359 "nop_in_interval": 30, 00:04:25.359 "disable_chap": false, 00:04:25.359 "require_chap": false, 00:04:25.359 "mutual_chap": false, 00:04:25.359 "chap_group": 0, 00:04:25.359 "max_large_datain_per_connection": 64, 00:04:25.359 "max_r2t_per_connection": 4, 00:04:25.359 "pdu_pool_size": 36864, 00:04:25.359 "immediate_data_pool_size": 16384, 00:04:25.359 "data_out_pool_size": 2048 00:04:25.359 } 00:04:25.359 } 00:04:25.359 ] 00:04:25.359 } 00:04:25.359 ] 00:04:25.359 } 00:04:25.359 17:50:19 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:04:25.359 17:50:19 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@40 -- # killprocess 405783 00:04:25.359 17:50:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@948 -- # '[' -z 405783 ']' 00:04:25.359 17:50:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@952 -- # kill -0 405783 00:04:25.359 17:50:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@953 -- # uname 00:04:25.619 17:50:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:04:25.619 17:50:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 405783 00:04:25.619 17:50:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:04:25.619 17:50:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:04:25.619 17:50:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@966 -- # echo 'killing process with pid 405783' 00:04:25.619 killing process with pid 405783 00:04:25.619 17:50:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@967 -- # kill 405783 00:04:25.619 17:50:19 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # wait 405783 00:04:25.878 17:50:19 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@47 -- # local spdk_pid=406005 00:04:25.878 17:50:19 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@48 -- # sleep 5 00:04:25.878 17:50:19 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:04:31.153 17:50:24 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@50 -- # killprocess 406005 00:04:31.153 17:50:24 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@948 -- # '[' -z 406005 ']' 00:04:31.153 17:50:24 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@952 -- # kill -0 406005 00:04:31.153 17:50:24 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@953 -- # uname 00:04:31.153 17:50:24 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:04:31.153 17:50:24 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 406005 00:04:31.153 17:50:24 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:04:31.153 17:50:24 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:04:31.153 17:50:24 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@966 -- # echo 'killing process with pid 406005' 00:04:31.153 killing process with pid 406005 00:04:31.153 17:50:24 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@967 -- # kill 406005 00:04:31.153 17:50:24 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@972 -- # wait 406005 00:04:31.153 17:50:24 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@51 -- # grep -q 'TCP Transport Init' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/log.txt 00:04:31.153 17:50:24 skip_rpc.skip_rpc_with_json -- rpc/skip_rpc.sh@52 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/log.txt 00:04:31.153 00:04:31.153 real 0m6.248s 00:04:31.153 user 0m5.982s 00:04:31.153 sys 0m0.540s 00:04:31.153 17:50:24 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:31.153 17:50:24 skip_rpc.skip_rpc_with_json -- common/autotest_common.sh@10 -- # set +x 00:04:31.153 ************************************ 00:04:31.153 END TEST skip_rpc_with_json 00:04:31.153 ************************************ 00:04:31.153 17:50:24 skip_rpc -- common/autotest_common.sh@1142 -- # return 0 00:04:31.153 17:50:24 skip_rpc -- rpc/skip_rpc.sh@75 -- # run_test skip_rpc_with_delay test_skip_rpc_with_delay 00:04:31.153 17:50:24 skip_rpc -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:31.153 17:50:24 skip_rpc -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:31.153 17:50:24 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:31.153 ************************************ 00:04:31.153 START TEST skip_rpc_with_delay 00:04:31.153 ************************************ 00:04:31.154 17:50:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1123 -- # test_skip_rpc_with_delay 00:04:31.154 17:50:24 skip_rpc.skip_rpc_with_delay -- rpc/skip_rpc.sh@57 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:04:31.154 17:50:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@648 -- # local es=0 00:04:31.154 17:50:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:04:31.154 17:50:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:31.154 17:50:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:04:31.154 17:50:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:31.154 17:50:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:04:31.154 17:50:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:31.154 17:50:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:04:31.154 17:50:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:31.154 17:50:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt ]] 00:04:31.154 17:50:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --no-rpc-server -m 0x1 --wait-for-rpc 00:04:31.412 [2024-07-15 17:50:24.897618] app.c: 832:spdk_app_start: *ERROR*: Cannot use '--wait-for-rpc' if no RPC server is going to be started. 00:04:31.412 [2024-07-15 17:50:24.897678] app.c: 711:unclaim_cpu_cores: *ERROR*: Failed to unlink lock fd for core 0, errno: 2 00:04:31.412 17:50:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@651 -- # es=1 00:04:31.412 17:50:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:04:31.412 17:50:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:04:31.412 17:50:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:04:31.412 00:04:31.412 real 0m0.066s 00:04:31.412 user 0m0.040s 00:04:31.412 sys 0m0.026s 00:04:31.412 17:50:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:31.412 17:50:24 skip_rpc.skip_rpc_with_delay -- common/autotest_common.sh@10 -- # set +x 00:04:31.412 ************************************ 00:04:31.412 END TEST skip_rpc_with_delay 00:04:31.412 ************************************ 00:04:31.412 17:50:24 skip_rpc -- common/autotest_common.sh@1142 -- # return 0 00:04:31.412 17:50:24 skip_rpc -- rpc/skip_rpc.sh@77 -- # uname 00:04:31.412 17:50:24 skip_rpc -- rpc/skip_rpc.sh@77 -- # '[' Linux '!=' FreeBSD ']' 00:04:31.412 17:50:24 skip_rpc -- rpc/skip_rpc.sh@78 -- # run_test exit_on_failed_rpc_init test_exit_on_failed_rpc_init 00:04:31.412 17:50:24 skip_rpc -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:31.412 17:50:24 skip_rpc -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:31.412 17:50:24 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:31.412 ************************************ 00:04:31.412 START TEST exit_on_failed_rpc_init 00:04:31.412 ************************************ 00:04:31.412 17:50:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1123 -- # test_exit_on_failed_rpc_init 00:04:31.412 17:50:24 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@62 -- # local spdk_pid=406988 00:04:31.412 17:50:24 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@63 -- # waitforlisten 406988 00:04:31.412 17:50:24 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:04:31.412 17:50:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@829 -- # '[' -z 406988 ']' 00:04:31.412 17:50:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:31.412 17:50:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@834 -- # local max_retries=100 00:04:31.412 17:50:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:31.412 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:31.412 17:50:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@838 -- # xtrace_disable 00:04:31.412 17:50:24 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:04:31.412 [2024-07-15 17:50:25.031821] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:04:31.413 [2024-07-15 17:50:25.031863] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid406988 ] 00:04:31.413 EAL: No free 2048 kB hugepages reported on node 1 00:04:31.413 [2024-07-15 17:50:25.084811] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:31.670 [2024-07-15 17:50:25.164236] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:04:32.235 17:50:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:04:32.235 17:50:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@862 -- # return 0 00:04:32.235 17:50:25 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@65 -- # trap 'killprocess $spdk_pid; exit 1' SIGINT SIGTERM EXIT 00:04:32.235 17:50:25 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@67 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:04:32.235 17:50:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@648 -- # local es=0 00:04:32.235 17:50:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:04:32.235 17:50:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:32.235 17:50:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:04:32.235 17:50:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:32.235 17:50:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:04:32.235 17:50:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:32.235 17:50:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:04:32.235 17:50:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:32.235 17:50:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt ]] 00:04:32.235 17:50:25 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x2 00:04:32.235 [2024-07-15 17:50:25.884557] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:04:32.235 [2024-07-15 17:50:25.884605] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid407010 ] 00:04:32.235 EAL: No free 2048 kB hugepages reported on node 1 00:04:32.235 [2024-07-15 17:50:25.940091] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:32.493 [2024-07-15 17:50:26.013579] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:04:32.493 [2024-07-15 17:50:26.013645] rpc.c: 180:_spdk_rpc_listen: *ERROR*: RPC Unix domain socket path /var/tmp/spdk.sock in use. Specify another. 00:04:32.493 [2024-07-15 17:50:26.013654] rpc.c: 166:spdk_rpc_initialize: *ERROR*: Unable to start RPC service at /var/tmp/spdk.sock 00:04:32.493 [2024-07-15 17:50:26.013660] app.c:1053:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:04:32.493 17:50:26 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@651 -- # es=234 00:04:32.493 17:50:26 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:04:32.493 17:50:26 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@660 -- # es=106 00:04:32.493 17:50:26 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@661 -- # case "$es" in 00:04:32.493 17:50:26 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@668 -- # es=1 00:04:32.493 17:50:26 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:04:32.493 17:50:26 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:04:32.493 17:50:26 skip_rpc.exit_on_failed_rpc_init -- rpc/skip_rpc.sh@70 -- # killprocess 406988 00:04:32.493 17:50:26 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@948 -- # '[' -z 406988 ']' 00:04:32.493 17:50:26 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@952 -- # kill -0 406988 00:04:32.493 17:50:26 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@953 -- # uname 00:04:32.493 17:50:26 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:04:32.493 17:50:26 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 406988 00:04:32.493 17:50:26 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:04:32.493 17:50:26 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:04:32.493 17:50:26 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@966 -- # echo 'killing process with pid 406988' 00:04:32.493 killing process with pid 406988 00:04:32.493 17:50:26 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@967 -- # kill 406988 00:04:32.493 17:50:26 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@972 -- # wait 406988 00:04:32.766 00:04:32.766 real 0m1.457s 00:04:32.766 user 0m1.693s 00:04:32.766 sys 0m0.387s 00:04:32.766 17:50:26 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:32.766 17:50:26 skip_rpc.exit_on_failed_rpc_init -- common/autotest_common.sh@10 -- # set +x 00:04:32.766 ************************************ 00:04:32.766 END TEST exit_on_failed_rpc_init 00:04:32.766 ************************************ 00:04:32.766 17:50:26 skip_rpc -- common/autotest_common.sh@1142 -- # return 0 00:04:32.766 17:50:26 skip_rpc -- rpc/skip_rpc.sh@81 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc/config.json 00:04:32.766 00:04:32.766 real 0m13.499s 00:04:32.766 user 0m13.005s 00:04:32.766 sys 0m1.443s 00:04:32.766 17:50:26 skip_rpc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:32.766 17:50:26 skip_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:32.766 ************************************ 00:04:32.766 END TEST skip_rpc 00:04:32.766 ************************************ 00:04:33.024 17:50:26 -- common/autotest_common.sh@1142 -- # return 0 00:04:33.024 17:50:26 -- spdk/autotest.sh@171 -- # run_test rpc_client /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client/rpc_client.sh 00:04:33.024 17:50:26 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:33.025 17:50:26 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:33.025 17:50:26 -- common/autotest_common.sh@10 -- # set +x 00:04:33.025 ************************************ 00:04:33.025 START TEST rpc_client 00:04:33.025 ************************************ 00:04:33.025 17:50:26 rpc_client -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client/rpc_client.sh 00:04:33.025 * Looking for test storage... 00:04:33.025 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client 00:04:33.025 17:50:26 rpc_client -- rpc_client/rpc_client.sh@10 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_client/rpc_client_test 00:04:33.025 OK 00:04:33.025 17:50:26 rpc_client -- rpc_client/rpc_client.sh@12 -- # trap - SIGINT SIGTERM EXIT 00:04:33.025 00:04:33.025 real 0m0.108s 00:04:33.025 user 0m0.048s 00:04:33.025 sys 0m0.067s 00:04:33.025 17:50:26 rpc_client -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:33.025 17:50:26 rpc_client -- common/autotest_common.sh@10 -- # set +x 00:04:33.025 ************************************ 00:04:33.025 END TEST rpc_client 00:04:33.025 ************************************ 00:04:33.025 17:50:26 -- common/autotest_common.sh@1142 -- # return 0 00:04:33.025 17:50:26 -- spdk/autotest.sh@172 -- # run_test json_config /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config.sh 00:04:33.025 17:50:26 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:33.025 17:50:26 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:33.025 17:50:26 -- common/autotest_common.sh@10 -- # set +x 00:04:33.025 ************************************ 00:04:33.025 START TEST json_config 00:04:33.025 ************************************ 00:04:33.025 17:50:26 json_config -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config.sh 00:04:33.285 17:50:26 json_config -- json_config/json_config.sh@8 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@7 -- # uname -s 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:04:33.285 17:50:26 json_config -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:04:33.285 17:50:26 json_config -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:04:33.285 17:50:26 json_config -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:04:33.285 17:50:26 json_config -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:33.285 17:50:26 json_config -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:33.285 17:50:26 json_config -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:33.285 17:50:26 json_config -- paths/export.sh@5 -- # export PATH 00:04:33.285 17:50:26 json_config -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@47 -- # : 0 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:04:33.285 17:50:26 json_config -- nvmf/common.sh@51 -- # have_pci_nics=0 00:04:33.285 17:50:26 json_config -- json_config/json_config.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/common.sh 00:04:33.285 17:50:26 json_config -- json_config/json_config.sh@11 -- # [[ 0 -eq 1 ]] 00:04:33.285 17:50:26 json_config -- json_config/json_config.sh@15 -- # [[ 0 -ne 1 ]] 00:04:33.285 17:50:26 json_config -- json_config/json_config.sh@15 -- # [[ 0 -eq 1 ]] 00:04:33.285 17:50:26 json_config -- json_config/json_config.sh@26 -- # (( SPDK_TEST_BLOCKDEV + SPDK_TEST_ISCSI + SPDK_TEST_NVMF + SPDK_TEST_VHOST + SPDK_TEST_VHOST_INIT + SPDK_TEST_RBD == 0 )) 00:04:33.285 17:50:26 json_config -- json_config/json_config.sh@31 -- # app_pid=(['target']='' ['initiator']='') 00:04:33.285 17:50:26 json_config -- json_config/json_config.sh@31 -- # declare -A app_pid 00:04:33.285 17:50:26 json_config -- json_config/json_config.sh@32 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock' ['initiator']='/var/tmp/spdk_initiator.sock') 00:04:33.286 17:50:26 json_config -- json_config/json_config.sh@32 -- # declare -A app_socket 00:04:33.286 17:50:26 json_config -- json_config/json_config.sh@33 -- # app_params=(['target']='-m 0x1 -s 1024' ['initiator']='-m 0x2 -g -u -s 1024') 00:04:33.286 17:50:26 json_config -- json_config/json_config.sh@33 -- # declare -A app_params 00:04:33.286 17:50:26 json_config -- json_config/json_config.sh@34 -- # configs_path=(['target']='/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json' ['initiator']='/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_initiator_config.json') 00:04:33.286 17:50:26 json_config -- json_config/json_config.sh@34 -- # declare -A configs_path 00:04:33.286 17:50:26 json_config -- json_config/json_config.sh@40 -- # last_event_id=0 00:04:33.286 17:50:26 json_config -- json_config/json_config.sh@355 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:04:33.286 17:50:26 json_config -- json_config/json_config.sh@356 -- # echo 'INFO: JSON configuration test init' 00:04:33.286 INFO: JSON configuration test init 00:04:33.286 17:50:26 json_config -- json_config/json_config.sh@357 -- # json_config_test_init 00:04:33.286 17:50:26 json_config -- json_config/json_config.sh@262 -- # timing_enter json_config_test_init 00:04:33.286 17:50:26 json_config -- common/autotest_common.sh@722 -- # xtrace_disable 00:04:33.286 17:50:26 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:33.286 17:50:26 json_config -- json_config/json_config.sh@263 -- # timing_enter json_config_setup_target 00:04:33.286 17:50:26 json_config -- common/autotest_common.sh@722 -- # xtrace_disable 00:04:33.286 17:50:26 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:33.286 17:50:26 json_config -- json_config/json_config.sh@265 -- # json_config_test_start_app target --wait-for-rpc 00:04:33.286 17:50:26 json_config -- json_config/common.sh@9 -- # local app=target 00:04:33.286 17:50:26 json_config -- json_config/common.sh@10 -- # shift 00:04:33.286 17:50:26 json_config -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:04:33.286 17:50:26 json_config -- json_config/common.sh@13 -- # [[ -z '' ]] 00:04:33.286 17:50:26 json_config -- json_config/common.sh@15 -- # local app_extra_params= 00:04:33.286 17:50:26 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:04:33.286 17:50:26 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:04:33.286 17:50:26 json_config -- json_config/common.sh@22 -- # app_pid["$app"]=407345 00:04:33.286 17:50:26 json_config -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:04:33.286 Waiting for target to run... 00:04:33.286 17:50:26 json_config -- json_config/common.sh@25 -- # waitforlisten 407345 /var/tmp/spdk_tgt.sock 00:04:33.286 17:50:26 json_config -- common/autotest_common.sh@829 -- # '[' -z 407345 ']' 00:04:33.286 17:50:26 json_config -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:04:33.286 17:50:26 json_config -- json_config/common.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --wait-for-rpc 00:04:33.286 17:50:26 json_config -- common/autotest_common.sh@834 -- # local max_retries=100 00:04:33.286 17:50:26 json_config -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:04:33.286 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:04:33.286 17:50:26 json_config -- common/autotest_common.sh@838 -- # xtrace_disable 00:04:33.286 17:50:26 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:33.286 [2024-07-15 17:50:26.865809] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:04:33.286 [2024-07-15 17:50:26.865861] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid407345 ] 00:04:33.286 EAL: No free 2048 kB hugepages reported on node 1 00:04:33.893 [2024-07-15 17:50:27.301425] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:33.893 [2024-07-15 17:50:27.387011] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:04:34.151 17:50:27 json_config -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:04:34.151 17:50:27 json_config -- common/autotest_common.sh@862 -- # return 0 00:04:34.151 17:50:27 json_config -- json_config/common.sh@26 -- # echo '' 00:04:34.151 00:04:34.151 17:50:27 json_config -- json_config/json_config.sh@269 -- # create_accel_config 00:04:34.151 17:50:27 json_config -- json_config/json_config.sh@93 -- # timing_enter create_accel_config 00:04:34.151 17:50:27 json_config -- common/autotest_common.sh@722 -- # xtrace_disable 00:04:34.151 17:50:27 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:34.151 17:50:27 json_config -- json_config/json_config.sh@95 -- # [[ 0 -eq 1 ]] 00:04:34.151 17:50:27 json_config -- json_config/json_config.sh@101 -- # timing_exit create_accel_config 00:04:34.151 17:50:27 json_config -- common/autotest_common.sh@728 -- # xtrace_disable 00:04:34.151 17:50:27 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:34.151 17:50:27 json_config -- json_config/json_config.sh@273 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh --json-with-subsystems 00:04:34.151 17:50:27 json_config -- json_config/json_config.sh@274 -- # tgt_rpc load_config 00:04:34.151 17:50:27 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock load_config 00:04:37.438 17:50:30 json_config -- json_config/json_config.sh@276 -- # tgt_check_notification_types 00:04:37.438 17:50:30 json_config -- json_config/json_config.sh@43 -- # timing_enter tgt_check_notification_types 00:04:37.438 17:50:30 json_config -- common/autotest_common.sh@722 -- # xtrace_disable 00:04:37.438 17:50:30 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:37.438 17:50:30 json_config -- json_config/json_config.sh@45 -- # local ret=0 00:04:37.438 17:50:30 json_config -- json_config/json_config.sh@46 -- # enabled_types=('bdev_register' 'bdev_unregister') 00:04:37.438 17:50:30 json_config -- json_config/json_config.sh@46 -- # local enabled_types 00:04:37.438 17:50:30 json_config -- json_config/json_config.sh@48 -- # tgt_rpc notify_get_types 00:04:37.438 17:50:30 json_config -- json_config/json_config.sh@48 -- # jq -r '.[]' 00:04:37.438 17:50:30 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock notify_get_types 00:04:37.438 17:50:30 json_config -- json_config/json_config.sh@48 -- # get_types=('bdev_register' 'bdev_unregister') 00:04:37.438 17:50:30 json_config -- json_config/json_config.sh@48 -- # local get_types 00:04:37.438 17:50:30 json_config -- json_config/json_config.sh@49 -- # [[ bdev_register bdev_unregister != \b\d\e\v\_\r\e\g\i\s\t\e\r\ \b\d\e\v\_\u\n\r\e\g\i\s\t\e\r ]] 00:04:37.438 17:50:30 json_config -- json_config/json_config.sh@54 -- # timing_exit tgt_check_notification_types 00:04:37.438 17:50:30 json_config -- common/autotest_common.sh@728 -- # xtrace_disable 00:04:37.439 17:50:30 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:37.439 17:50:30 json_config -- json_config/json_config.sh@55 -- # return 0 00:04:37.439 17:50:30 json_config -- json_config/json_config.sh@278 -- # [[ 0 -eq 1 ]] 00:04:37.439 17:50:30 json_config -- json_config/json_config.sh@282 -- # [[ 0 -eq 1 ]] 00:04:37.439 17:50:30 json_config -- json_config/json_config.sh@286 -- # [[ 0 -eq 1 ]] 00:04:37.439 17:50:30 json_config -- json_config/json_config.sh@290 -- # [[ 1 -eq 1 ]] 00:04:37.439 17:50:30 json_config -- json_config/json_config.sh@291 -- # create_nvmf_subsystem_config 00:04:37.439 17:50:30 json_config -- json_config/json_config.sh@230 -- # timing_enter create_nvmf_subsystem_config 00:04:37.439 17:50:30 json_config -- common/autotest_common.sh@722 -- # xtrace_disable 00:04:37.439 17:50:30 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:37.439 17:50:30 json_config -- json_config/json_config.sh@232 -- # NVMF_FIRST_TARGET_IP=127.0.0.1 00:04:37.439 17:50:30 json_config -- json_config/json_config.sh@233 -- # [[ tcp == \r\d\m\a ]] 00:04:37.439 17:50:30 json_config -- json_config/json_config.sh@237 -- # [[ -z 127.0.0.1 ]] 00:04:37.439 17:50:30 json_config -- json_config/json_config.sh@242 -- # tgt_rpc bdev_malloc_create 8 512 --name MallocForNvmf0 00:04:37.439 17:50:30 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 8 512 --name MallocForNvmf0 00:04:37.439 MallocForNvmf0 00:04:37.439 17:50:31 json_config -- json_config/json_config.sh@243 -- # tgt_rpc bdev_malloc_create 4 1024 --name MallocForNvmf1 00:04:37.439 17:50:31 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 4 1024 --name MallocForNvmf1 00:04:37.698 MallocForNvmf1 00:04:37.698 17:50:31 json_config -- json_config/json_config.sh@245 -- # tgt_rpc nvmf_create_transport -t tcp -u 8192 -c 0 00:04:37.698 17:50:31 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_create_transport -t tcp -u 8192 -c 0 00:04:37.957 [2024-07-15 17:50:31.480828] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:04:37.957 17:50:31 json_config -- json_config/json_config.sh@246 -- # tgt_rpc nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:04:37.957 17:50:31 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:04:38.215 17:50:31 json_config -- json_config/json_config.sh@247 -- # tgt_rpc nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf0 00:04:38.215 17:50:31 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf0 00:04:38.215 17:50:31 json_config -- json_config/json_config.sh@248 -- # tgt_rpc nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf1 00:04:38.215 17:50:31 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 MallocForNvmf1 00:04:38.473 17:50:32 json_config -- json_config/json_config.sh@249 -- # tgt_rpc nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 127.0.0.1 -s 4420 00:04:38.473 17:50:32 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 127.0.0.1 -s 4420 00:04:38.473 [2024-07-15 17:50:32.186992] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:04:38.732 17:50:32 json_config -- json_config/json_config.sh@251 -- # timing_exit create_nvmf_subsystem_config 00:04:38.732 17:50:32 json_config -- common/autotest_common.sh@728 -- # xtrace_disable 00:04:38.732 17:50:32 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:38.732 17:50:32 json_config -- json_config/json_config.sh@293 -- # timing_exit json_config_setup_target 00:04:38.732 17:50:32 json_config -- common/autotest_common.sh@728 -- # xtrace_disable 00:04:38.732 17:50:32 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:38.732 17:50:32 json_config -- json_config/json_config.sh@295 -- # [[ 0 -eq 1 ]] 00:04:38.732 17:50:32 json_config -- json_config/json_config.sh@300 -- # tgt_rpc bdev_malloc_create 8 512 --name MallocBdevForConfigChangeCheck 00:04:38.732 17:50:32 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_create 8 512 --name MallocBdevForConfigChangeCheck 00:04:38.732 MallocBdevForConfigChangeCheck 00:04:38.732 17:50:32 json_config -- json_config/json_config.sh@302 -- # timing_exit json_config_test_init 00:04:38.732 17:50:32 json_config -- common/autotest_common.sh@728 -- # xtrace_disable 00:04:38.732 17:50:32 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:38.991 17:50:32 json_config -- json_config/json_config.sh@359 -- # tgt_rpc save_config 00:04:38.991 17:50:32 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:04:39.251 17:50:32 json_config -- json_config/json_config.sh@361 -- # echo 'INFO: shutting down applications...' 00:04:39.251 INFO: shutting down applications... 00:04:39.251 17:50:32 json_config -- json_config/json_config.sh@362 -- # [[ 0 -eq 1 ]] 00:04:39.251 17:50:32 json_config -- json_config/json_config.sh@368 -- # json_config_clear target 00:04:39.251 17:50:32 json_config -- json_config/json_config.sh@332 -- # [[ -n 22 ]] 00:04:39.251 17:50:32 json_config -- json_config/json_config.sh@333 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/clear_config.py -s /var/tmp/spdk_tgt.sock clear_config 00:04:40.650 Calling clear_iscsi_subsystem 00:04:40.651 Calling clear_nvmf_subsystem 00:04:40.651 Calling clear_nbd_subsystem 00:04:40.651 Calling clear_ublk_subsystem 00:04:40.651 Calling clear_vhost_blk_subsystem 00:04:40.651 Calling clear_vhost_scsi_subsystem 00:04:40.651 Calling clear_bdev_subsystem 00:04:40.651 17:50:34 json_config -- json_config/json_config.sh@337 -- # local config_filter=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py 00:04:40.651 17:50:34 json_config -- json_config/json_config.sh@343 -- # count=100 00:04:40.651 17:50:34 json_config -- json_config/json_config.sh@344 -- # '[' 100 -gt 0 ']' 00:04:40.651 17:50:34 json_config -- json_config/json_config.sh@345 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method check_empty 00:04:40.651 17:50:34 json_config -- json_config/json_config.sh@345 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:04:40.651 17:50:34 json_config -- json_config/json_config.sh@345 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method delete_global_parameters 00:04:41.219 17:50:34 json_config -- json_config/json_config.sh@345 -- # break 00:04:41.219 17:50:34 json_config -- json_config/json_config.sh@350 -- # '[' 100 -eq 0 ']' 00:04:41.219 17:50:34 json_config -- json_config/json_config.sh@369 -- # json_config_test_shutdown_app target 00:04:41.219 17:50:34 json_config -- json_config/common.sh@31 -- # local app=target 00:04:41.219 17:50:34 json_config -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:04:41.219 17:50:34 json_config -- json_config/common.sh@35 -- # [[ -n 407345 ]] 00:04:41.219 17:50:34 json_config -- json_config/common.sh@38 -- # kill -SIGINT 407345 00:04:41.219 17:50:34 json_config -- json_config/common.sh@40 -- # (( i = 0 )) 00:04:41.219 17:50:34 json_config -- json_config/common.sh@40 -- # (( i < 30 )) 00:04:41.219 17:50:34 json_config -- json_config/common.sh@41 -- # kill -0 407345 00:04:41.219 17:50:34 json_config -- json_config/common.sh@45 -- # sleep 0.5 00:04:41.478 17:50:35 json_config -- json_config/common.sh@40 -- # (( i++ )) 00:04:41.478 17:50:35 json_config -- json_config/common.sh@40 -- # (( i < 30 )) 00:04:41.478 17:50:35 json_config -- json_config/common.sh@41 -- # kill -0 407345 00:04:41.478 17:50:35 json_config -- json_config/common.sh@42 -- # app_pid["$app"]= 00:04:41.478 17:50:35 json_config -- json_config/common.sh@43 -- # break 00:04:41.478 17:50:35 json_config -- json_config/common.sh@48 -- # [[ -n '' ]] 00:04:41.478 17:50:35 json_config -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:04:41.478 SPDK target shutdown done 00:04:41.478 17:50:35 json_config -- json_config/json_config.sh@371 -- # echo 'INFO: relaunching applications...' 00:04:41.478 INFO: relaunching applications... 00:04:41.478 17:50:35 json_config -- json_config/json_config.sh@372 -- # json_config_test_start_app target --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:04:41.478 17:50:35 json_config -- json_config/common.sh@9 -- # local app=target 00:04:41.478 17:50:35 json_config -- json_config/common.sh@10 -- # shift 00:04:41.478 17:50:35 json_config -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:04:41.478 17:50:35 json_config -- json_config/common.sh@13 -- # [[ -z '' ]] 00:04:41.478 17:50:35 json_config -- json_config/common.sh@15 -- # local app_extra_params= 00:04:41.478 17:50:35 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:04:41.478 17:50:35 json_config -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:04:41.478 17:50:35 json_config -- json_config/common.sh@22 -- # app_pid["$app"]=408850 00:04:41.478 17:50:35 json_config -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:04:41.478 Waiting for target to run... 00:04:41.478 17:50:35 json_config -- json_config/common.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:04:41.478 17:50:35 json_config -- json_config/common.sh@25 -- # waitforlisten 408850 /var/tmp/spdk_tgt.sock 00:04:41.478 17:50:35 json_config -- common/autotest_common.sh@829 -- # '[' -z 408850 ']' 00:04:41.478 17:50:35 json_config -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:04:41.478 17:50:35 json_config -- common/autotest_common.sh@834 -- # local max_retries=100 00:04:41.478 17:50:35 json_config -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:04:41.478 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:04:41.478 17:50:35 json_config -- common/autotest_common.sh@838 -- # xtrace_disable 00:04:41.478 17:50:35 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:41.738 [2024-07-15 17:50:35.206793] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:04:41.738 [2024-07-15 17:50:35.206846] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid408850 ] 00:04:41.738 EAL: No free 2048 kB hugepages reported on node 1 00:04:41.997 [2024-07-15 17:50:35.468392] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:41.997 [2024-07-15 17:50:35.539893] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:04:45.288 [2024-07-15 17:50:38.550677] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:04:45.288 [2024-07-15 17:50:38.582981] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:04:45.288 17:50:38 json_config -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:04:45.289 17:50:38 json_config -- common/autotest_common.sh@862 -- # return 0 00:04:45.289 17:50:38 json_config -- json_config/common.sh@26 -- # echo '' 00:04:45.289 00:04:45.289 17:50:38 json_config -- json_config/json_config.sh@373 -- # [[ 0 -eq 1 ]] 00:04:45.289 17:50:38 json_config -- json_config/json_config.sh@377 -- # echo 'INFO: Checking if target configuration is the same...' 00:04:45.289 INFO: Checking if target configuration is the same... 00:04:45.289 17:50:38 json_config -- json_config/json_config.sh@378 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh /dev/fd/62 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:04:45.289 17:50:38 json_config -- json_config/json_config.sh@378 -- # tgt_rpc save_config 00:04:45.289 17:50:38 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:04:45.289 + '[' 2 -ne 2 ']' 00:04:45.289 +++ dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh 00:04:45.289 ++ readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/../.. 00:04:45.289 + rootdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:04:45.289 +++ basename /dev/fd/62 00:04:45.289 ++ mktemp /tmp/62.XXX 00:04:45.289 + tmp_file_1=/tmp/62.tZM 00:04:45.289 +++ basename /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:04:45.289 ++ mktemp /tmp/spdk_tgt_config.json.XXX 00:04:45.289 + tmp_file_2=/tmp/spdk_tgt_config.json.uCM 00:04:45.289 + ret=0 00:04:45.289 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:04:45.289 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:04:45.289 + diff -u /tmp/62.tZM /tmp/spdk_tgt_config.json.uCM 00:04:45.289 + echo 'INFO: JSON config files are the same' 00:04:45.289 INFO: JSON config files are the same 00:04:45.289 + rm /tmp/62.tZM /tmp/spdk_tgt_config.json.uCM 00:04:45.289 + exit 0 00:04:45.289 17:50:38 json_config -- json_config/json_config.sh@379 -- # [[ 0 -eq 1 ]] 00:04:45.289 17:50:38 json_config -- json_config/json_config.sh@384 -- # echo 'INFO: changing configuration and checking if this can be detected...' 00:04:45.289 INFO: changing configuration and checking if this can be detected... 00:04:45.289 17:50:38 json_config -- json_config/json_config.sh@386 -- # tgt_rpc bdev_malloc_delete MallocBdevForConfigChangeCheck 00:04:45.289 17:50:38 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock bdev_malloc_delete MallocBdevForConfigChangeCheck 00:04:45.548 17:50:39 json_config -- json_config/json_config.sh@387 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh /dev/fd/62 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:04:45.548 17:50:39 json_config -- json_config/json_config.sh@387 -- # tgt_rpc save_config 00:04:45.548 17:50:39 json_config -- json_config/common.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk_tgt.sock save_config 00:04:45.548 + '[' 2 -ne 2 ']' 00:04:45.548 +++ dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_diff.sh 00:04:45.548 ++ readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/../.. 00:04:45.548 + rootdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:04:45.548 +++ basename /dev/fd/62 00:04:45.548 ++ mktemp /tmp/62.XXX 00:04:45.548 + tmp_file_1=/tmp/62.v5i 00:04:45.548 +++ basename /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:04:45.548 ++ mktemp /tmp/spdk_tgt_config.json.XXX 00:04:45.548 + tmp_file_2=/tmp/spdk_tgt_config.json.d08 00:04:45.548 + ret=0 00:04:45.548 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:04:45.807 + /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/config_filter.py -method sort 00:04:45.807 + diff -u /tmp/62.v5i /tmp/spdk_tgt_config.json.d08 00:04:45.807 + ret=1 00:04:45.807 + echo '=== Start of file: /tmp/62.v5i ===' 00:04:45.807 + cat /tmp/62.v5i 00:04:45.807 + echo '=== End of file: /tmp/62.v5i ===' 00:04:45.807 + echo '' 00:04:45.807 + echo '=== Start of file: /tmp/spdk_tgt_config.json.d08 ===' 00:04:45.807 + cat /tmp/spdk_tgt_config.json.d08 00:04:45.807 + echo '=== End of file: /tmp/spdk_tgt_config.json.d08 ===' 00:04:45.807 + echo '' 00:04:45.807 + rm /tmp/62.v5i /tmp/spdk_tgt_config.json.d08 00:04:45.807 + exit 1 00:04:45.807 17:50:39 json_config -- json_config/json_config.sh@391 -- # echo 'INFO: configuration change detected.' 00:04:45.807 INFO: configuration change detected. 00:04:45.807 17:50:39 json_config -- json_config/json_config.sh@394 -- # json_config_test_fini 00:04:45.807 17:50:39 json_config -- json_config/json_config.sh@306 -- # timing_enter json_config_test_fini 00:04:45.807 17:50:39 json_config -- common/autotest_common.sh@722 -- # xtrace_disable 00:04:45.807 17:50:39 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:45.807 17:50:39 json_config -- json_config/json_config.sh@307 -- # local ret=0 00:04:45.807 17:50:39 json_config -- json_config/json_config.sh@309 -- # [[ -n '' ]] 00:04:45.807 17:50:39 json_config -- json_config/json_config.sh@317 -- # [[ -n 408850 ]] 00:04:45.807 17:50:39 json_config -- json_config/json_config.sh@320 -- # cleanup_bdev_subsystem_config 00:04:45.807 17:50:39 json_config -- json_config/json_config.sh@184 -- # timing_enter cleanup_bdev_subsystem_config 00:04:45.807 17:50:39 json_config -- common/autotest_common.sh@722 -- # xtrace_disable 00:04:45.807 17:50:39 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:45.807 17:50:39 json_config -- json_config/json_config.sh@186 -- # [[ 0 -eq 1 ]] 00:04:45.807 17:50:39 json_config -- json_config/json_config.sh@193 -- # uname -s 00:04:45.807 17:50:39 json_config -- json_config/json_config.sh@193 -- # [[ Linux = Linux ]] 00:04:45.807 17:50:39 json_config -- json_config/json_config.sh@194 -- # rm -f /sample_aio 00:04:45.807 17:50:39 json_config -- json_config/json_config.sh@197 -- # [[ 0 -eq 1 ]] 00:04:45.807 17:50:39 json_config -- json_config/json_config.sh@201 -- # timing_exit cleanup_bdev_subsystem_config 00:04:45.807 17:50:39 json_config -- common/autotest_common.sh@728 -- # xtrace_disable 00:04:45.807 17:50:39 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:46.067 17:50:39 json_config -- json_config/json_config.sh@323 -- # killprocess 408850 00:04:46.067 17:50:39 json_config -- common/autotest_common.sh@948 -- # '[' -z 408850 ']' 00:04:46.067 17:50:39 json_config -- common/autotest_common.sh@952 -- # kill -0 408850 00:04:46.067 17:50:39 json_config -- common/autotest_common.sh@953 -- # uname 00:04:46.067 17:50:39 json_config -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:04:46.067 17:50:39 json_config -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 408850 00:04:46.067 17:50:39 json_config -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:04:46.067 17:50:39 json_config -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:04:46.067 17:50:39 json_config -- common/autotest_common.sh@966 -- # echo 'killing process with pid 408850' 00:04:46.067 killing process with pid 408850 00:04:46.067 17:50:39 json_config -- common/autotest_common.sh@967 -- # kill 408850 00:04:46.067 17:50:39 json_config -- common/autotest_common.sh@972 -- # wait 408850 00:04:47.445 17:50:41 json_config -- json_config/json_config.sh@326 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_initiator_config.json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/spdk_tgt_config.json 00:04:47.445 17:50:41 json_config -- json_config/json_config.sh@327 -- # timing_exit json_config_test_fini 00:04:47.445 17:50:41 json_config -- common/autotest_common.sh@728 -- # xtrace_disable 00:04:47.445 17:50:41 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:47.445 17:50:41 json_config -- json_config/json_config.sh@328 -- # return 0 00:04:47.445 17:50:41 json_config -- json_config/json_config.sh@396 -- # echo 'INFO: Success' 00:04:47.445 INFO: Success 00:04:47.445 00:04:47.445 real 0m14.375s 00:04:47.445 user 0m14.922s 00:04:47.445 sys 0m1.844s 00:04:47.445 17:50:41 json_config -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:47.445 17:50:41 json_config -- common/autotest_common.sh@10 -- # set +x 00:04:47.445 ************************************ 00:04:47.445 END TEST json_config 00:04:47.445 ************************************ 00:04:47.445 17:50:41 -- common/autotest_common.sh@1142 -- # return 0 00:04:47.445 17:50:41 -- spdk/autotest.sh@173 -- # run_test json_config_extra_key /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config_extra_key.sh 00:04:47.445 17:50:41 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:47.445 17:50:41 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:47.445 17:50:41 -- common/autotest_common.sh@10 -- # set +x 00:04:47.445 ************************************ 00:04:47.445 START TEST json_config_extra_key 00:04:47.445 ************************************ 00:04:47.445 17:50:41 json_config_extra_key -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/json_config_extra_key.sh 00:04:47.705 17:50:41 json_config_extra_key -- json_config/json_config_extra_key.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@7 -- # uname -s 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@21 -- # NET_TYPE=phy-fallback 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:04:47.705 17:50:41 json_config_extra_key -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:04:47.705 17:50:41 json_config_extra_key -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:04:47.705 17:50:41 json_config_extra_key -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:04:47.705 17:50:41 json_config_extra_key -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:47.705 17:50:41 json_config_extra_key -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:47.705 17:50:41 json_config_extra_key -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:47.705 17:50:41 json_config_extra_key -- paths/export.sh@5 -- # export PATH 00:04:47.705 17:50:41 json_config_extra_key -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@47 -- # : 0 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:04:47.705 17:50:41 json_config_extra_key -- nvmf/common.sh@51 -- # have_pci_nics=0 00:04:47.705 17:50:41 json_config_extra_key -- json_config/json_config_extra_key.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/common.sh 00:04:47.705 17:50:41 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # app_pid=(['target']='') 00:04:47.705 17:50:41 json_config_extra_key -- json_config/json_config_extra_key.sh@17 -- # declare -A app_pid 00:04:47.705 17:50:41 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # app_socket=(['target']='/var/tmp/spdk_tgt.sock') 00:04:47.705 17:50:41 json_config_extra_key -- json_config/json_config_extra_key.sh@18 -- # declare -A app_socket 00:04:47.705 17:50:41 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # app_params=(['target']='-m 0x1 -s 1024') 00:04:47.705 17:50:41 json_config_extra_key -- json_config/json_config_extra_key.sh@19 -- # declare -A app_params 00:04:47.705 17:50:41 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # configs_path=(['target']='/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/extra_key.json') 00:04:47.705 17:50:41 json_config_extra_key -- json_config/json_config_extra_key.sh@20 -- # declare -A configs_path 00:04:47.705 17:50:41 json_config_extra_key -- json_config/json_config_extra_key.sh@22 -- # trap 'on_error_exit "${FUNCNAME}" "${LINENO}"' ERR 00:04:47.705 17:50:41 json_config_extra_key -- json_config/json_config_extra_key.sh@24 -- # echo 'INFO: launching applications...' 00:04:47.705 INFO: launching applications... 00:04:47.705 17:50:41 json_config_extra_key -- json_config/json_config_extra_key.sh@25 -- # json_config_test_start_app target --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/extra_key.json 00:04:47.705 17:50:41 json_config_extra_key -- json_config/common.sh@9 -- # local app=target 00:04:47.705 17:50:41 json_config_extra_key -- json_config/common.sh@10 -- # shift 00:04:47.705 17:50:41 json_config_extra_key -- json_config/common.sh@12 -- # [[ -n 22 ]] 00:04:47.705 17:50:41 json_config_extra_key -- json_config/common.sh@13 -- # [[ -z '' ]] 00:04:47.705 17:50:41 json_config_extra_key -- json_config/common.sh@15 -- # local app_extra_params= 00:04:47.705 17:50:41 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:04:47.705 17:50:41 json_config_extra_key -- json_config/common.sh@16 -- # [[ 0 -eq 1 ]] 00:04:47.705 17:50:41 json_config_extra_key -- json_config/common.sh@22 -- # app_pid["$app"]=410120 00:04:47.705 17:50:41 json_config_extra_key -- json_config/common.sh@24 -- # echo 'Waiting for target to run...' 00:04:47.705 Waiting for target to run... 00:04:47.705 17:50:41 json_config_extra_key -- json_config/common.sh@25 -- # waitforlisten 410120 /var/tmp/spdk_tgt.sock 00:04:47.705 17:50:41 json_config_extra_key -- json_config/common.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -s 1024 -r /var/tmp/spdk_tgt.sock --json /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/extra_key.json 00:04:47.706 17:50:41 json_config_extra_key -- common/autotest_common.sh@829 -- # '[' -z 410120 ']' 00:04:47.706 17:50:41 json_config_extra_key -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk_tgt.sock 00:04:47.706 17:50:41 json_config_extra_key -- common/autotest_common.sh@834 -- # local max_retries=100 00:04:47.706 17:50:41 json_config_extra_key -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock...' 00:04:47.706 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk_tgt.sock... 00:04:47.706 17:50:41 json_config_extra_key -- common/autotest_common.sh@838 -- # xtrace_disable 00:04:47.706 17:50:41 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:04:47.706 [2024-07-15 17:50:41.299818] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:04:47.706 [2024-07-15 17:50:41.299866] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 -m 1024 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid410120 ] 00:04:47.706 EAL: No free 2048 kB hugepages reported on node 1 00:04:48.274 [2024-07-15 17:50:41.731139] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:48.274 [2024-07-15 17:50:41.816145] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:04:48.533 17:50:42 json_config_extra_key -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:04:48.533 17:50:42 json_config_extra_key -- common/autotest_common.sh@862 -- # return 0 00:04:48.533 17:50:42 json_config_extra_key -- json_config/common.sh@26 -- # echo '' 00:04:48.533 00:04:48.533 17:50:42 json_config_extra_key -- json_config/json_config_extra_key.sh@27 -- # echo 'INFO: shutting down applications...' 00:04:48.533 INFO: shutting down applications... 00:04:48.533 17:50:42 json_config_extra_key -- json_config/json_config_extra_key.sh@28 -- # json_config_test_shutdown_app target 00:04:48.533 17:50:42 json_config_extra_key -- json_config/common.sh@31 -- # local app=target 00:04:48.533 17:50:42 json_config_extra_key -- json_config/common.sh@34 -- # [[ -n 22 ]] 00:04:48.533 17:50:42 json_config_extra_key -- json_config/common.sh@35 -- # [[ -n 410120 ]] 00:04:48.533 17:50:42 json_config_extra_key -- json_config/common.sh@38 -- # kill -SIGINT 410120 00:04:48.533 17:50:42 json_config_extra_key -- json_config/common.sh@40 -- # (( i = 0 )) 00:04:48.533 17:50:42 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:04:48.533 17:50:42 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 410120 00:04:48.533 17:50:42 json_config_extra_key -- json_config/common.sh@45 -- # sleep 0.5 00:04:49.102 17:50:42 json_config_extra_key -- json_config/common.sh@40 -- # (( i++ )) 00:04:49.102 17:50:42 json_config_extra_key -- json_config/common.sh@40 -- # (( i < 30 )) 00:04:49.102 17:50:42 json_config_extra_key -- json_config/common.sh@41 -- # kill -0 410120 00:04:49.102 17:50:42 json_config_extra_key -- json_config/common.sh@42 -- # app_pid["$app"]= 00:04:49.102 17:50:42 json_config_extra_key -- json_config/common.sh@43 -- # break 00:04:49.102 17:50:42 json_config_extra_key -- json_config/common.sh@48 -- # [[ -n '' ]] 00:04:49.102 17:50:42 json_config_extra_key -- json_config/common.sh@53 -- # echo 'SPDK target shutdown done' 00:04:49.102 SPDK target shutdown done 00:04:49.102 17:50:42 json_config_extra_key -- json_config/json_config_extra_key.sh@30 -- # echo Success 00:04:49.102 Success 00:04:49.102 00:04:49.102 real 0m1.441s 00:04:49.102 user 0m1.051s 00:04:49.102 sys 0m0.537s 00:04:49.102 17:50:42 json_config_extra_key -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:49.102 17:50:42 json_config_extra_key -- common/autotest_common.sh@10 -- # set +x 00:04:49.102 ************************************ 00:04:49.102 END TEST json_config_extra_key 00:04:49.102 ************************************ 00:04:49.102 17:50:42 -- common/autotest_common.sh@1142 -- # return 0 00:04:49.102 17:50:42 -- spdk/autotest.sh@174 -- # run_test alias_rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:04:49.102 17:50:42 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:49.102 17:50:42 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:49.102 17:50:42 -- common/autotest_common.sh@10 -- # set +x 00:04:49.102 ************************************ 00:04:49.102 START TEST alias_rpc 00:04:49.102 ************************************ 00:04:49.102 17:50:42 alias_rpc -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/alias_rpc/alias_rpc.sh 00:04:49.102 * Looking for test storage... 00:04:49.102 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/alias_rpc 00:04:49.102 17:50:42 alias_rpc -- alias_rpc/alias_rpc.sh@10 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:04:49.102 17:50:42 alias_rpc -- alias_rpc/alias_rpc.sh@13 -- # spdk_tgt_pid=410397 00:04:49.102 17:50:42 alias_rpc -- alias_rpc/alias_rpc.sh@14 -- # waitforlisten 410397 00:04:49.102 17:50:42 alias_rpc -- alias_rpc/alias_rpc.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:49.102 17:50:42 alias_rpc -- common/autotest_common.sh@829 -- # '[' -z 410397 ']' 00:04:49.102 17:50:42 alias_rpc -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:49.102 17:50:42 alias_rpc -- common/autotest_common.sh@834 -- # local max_retries=100 00:04:49.102 17:50:42 alias_rpc -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:49.102 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:49.102 17:50:42 alias_rpc -- common/autotest_common.sh@838 -- # xtrace_disable 00:04:49.102 17:50:42 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:49.102 [2024-07-15 17:50:42.792862] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:04:49.102 [2024-07-15 17:50:42.792910] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid410397 ] 00:04:49.102 EAL: No free 2048 kB hugepages reported on node 1 00:04:49.361 [2024-07-15 17:50:42.845974] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:49.361 [2024-07-15 17:50:42.920863] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:04:49.929 17:50:43 alias_rpc -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:04:49.929 17:50:43 alias_rpc -- common/autotest_common.sh@862 -- # return 0 00:04:49.929 17:50:43 alias_rpc -- alias_rpc/alias_rpc.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py load_config -i 00:04:50.188 17:50:43 alias_rpc -- alias_rpc/alias_rpc.sh@19 -- # killprocess 410397 00:04:50.188 17:50:43 alias_rpc -- common/autotest_common.sh@948 -- # '[' -z 410397 ']' 00:04:50.188 17:50:43 alias_rpc -- common/autotest_common.sh@952 -- # kill -0 410397 00:04:50.188 17:50:43 alias_rpc -- common/autotest_common.sh@953 -- # uname 00:04:50.188 17:50:43 alias_rpc -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:04:50.188 17:50:43 alias_rpc -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 410397 00:04:50.188 17:50:43 alias_rpc -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:04:50.188 17:50:43 alias_rpc -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:04:50.188 17:50:43 alias_rpc -- common/autotest_common.sh@966 -- # echo 'killing process with pid 410397' 00:04:50.188 killing process with pid 410397 00:04:50.188 17:50:43 alias_rpc -- common/autotest_common.sh@967 -- # kill 410397 00:04:50.188 17:50:43 alias_rpc -- common/autotest_common.sh@972 -- # wait 410397 00:04:50.447 00:04:50.447 real 0m1.466s 00:04:50.447 user 0m1.624s 00:04:50.447 sys 0m0.375s 00:04:50.447 17:50:44 alias_rpc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:50.447 17:50:44 alias_rpc -- common/autotest_common.sh@10 -- # set +x 00:04:50.447 ************************************ 00:04:50.447 END TEST alias_rpc 00:04:50.447 ************************************ 00:04:50.447 17:50:44 -- common/autotest_common.sh@1142 -- # return 0 00:04:50.447 17:50:44 -- spdk/autotest.sh@176 -- # [[ 0 -eq 0 ]] 00:04:50.447 17:50:44 -- spdk/autotest.sh@177 -- # run_test spdkcli_tcp /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/tcp.sh 00:04:50.447 17:50:44 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:50.447 17:50:44 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:50.447 17:50:44 -- common/autotest_common.sh@10 -- # set +x 00:04:50.706 ************************************ 00:04:50.706 START TEST spdkcli_tcp 00:04:50.706 ************************************ 00:04:50.706 17:50:44 spdkcli_tcp -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/tcp.sh 00:04:50.706 * Looking for test storage... 00:04:50.706 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli 00:04:50.706 17:50:44 spdkcli_tcp -- spdkcli/tcp.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/common.sh 00:04:50.706 17:50:44 spdkcli_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py 00:04:50.706 17:50:44 spdkcli_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/clear_config.py 00:04:50.706 17:50:44 spdkcli_tcp -- spdkcli/tcp.sh@18 -- # IP_ADDRESS=127.0.0.1 00:04:50.706 17:50:44 spdkcli_tcp -- spdkcli/tcp.sh@19 -- # PORT=9998 00:04:50.706 17:50:44 spdkcli_tcp -- spdkcli/tcp.sh@21 -- # trap 'err_cleanup; exit 1' SIGINT SIGTERM EXIT 00:04:50.707 17:50:44 spdkcli_tcp -- spdkcli/tcp.sh@23 -- # timing_enter run_spdk_tgt_tcp 00:04:50.707 17:50:44 spdkcli_tcp -- common/autotest_common.sh@722 -- # xtrace_disable 00:04:50.707 17:50:44 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:04:50.707 17:50:44 spdkcli_tcp -- spdkcli/tcp.sh@25 -- # spdk_tgt_pid=410687 00:04:50.707 17:50:44 spdkcli_tcp -- spdkcli/tcp.sh@27 -- # waitforlisten 410687 00:04:50.707 17:50:44 spdkcli_tcp -- spdkcli/tcp.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x3 -p 0 00:04:50.707 17:50:44 spdkcli_tcp -- common/autotest_common.sh@829 -- # '[' -z 410687 ']' 00:04:50.707 17:50:44 spdkcli_tcp -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:50.707 17:50:44 spdkcli_tcp -- common/autotest_common.sh@834 -- # local max_retries=100 00:04:50.707 17:50:44 spdkcli_tcp -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:50.707 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:50.707 17:50:44 spdkcli_tcp -- common/autotest_common.sh@838 -- # xtrace_disable 00:04:50.707 17:50:44 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:04:50.707 [2024-07-15 17:50:44.344619] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:04:50.707 [2024-07-15 17:50:44.344662] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid410687 ] 00:04:50.707 EAL: No free 2048 kB hugepages reported on node 1 00:04:50.707 [2024-07-15 17:50:44.396876] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:04:50.965 [2024-07-15 17:50:44.470497] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:04:50.965 [2024-07-15 17:50:44.470500] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:04:51.534 17:50:45 spdkcli_tcp -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:04:51.534 17:50:45 spdkcli_tcp -- common/autotest_common.sh@862 -- # return 0 00:04:51.534 17:50:45 spdkcli_tcp -- spdkcli/tcp.sh@31 -- # socat_pid=410813 00:04:51.534 17:50:45 spdkcli_tcp -- spdkcli/tcp.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -r 100 -t 2 -s 127.0.0.1 -p 9998 rpc_get_methods 00:04:51.534 17:50:45 spdkcli_tcp -- spdkcli/tcp.sh@30 -- # socat TCP-LISTEN:9998 UNIX-CONNECT:/var/tmp/spdk.sock 00:04:51.822 [ 00:04:51.822 "bdev_malloc_delete", 00:04:51.822 "bdev_malloc_create", 00:04:51.822 "bdev_null_resize", 00:04:51.822 "bdev_null_delete", 00:04:51.822 "bdev_null_create", 00:04:51.822 "bdev_nvme_cuse_unregister", 00:04:51.822 "bdev_nvme_cuse_register", 00:04:51.822 "bdev_opal_new_user", 00:04:51.822 "bdev_opal_set_lock_state", 00:04:51.822 "bdev_opal_delete", 00:04:51.822 "bdev_opal_get_info", 00:04:51.822 "bdev_opal_create", 00:04:51.822 "bdev_nvme_opal_revert", 00:04:51.822 "bdev_nvme_opal_init", 00:04:51.822 "bdev_nvme_send_cmd", 00:04:51.822 "bdev_nvme_get_path_iostat", 00:04:51.822 "bdev_nvme_get_mdns_discovery_info", 00:04:51.822 "bdev_nvme_stop_mdns_discovery", 00:04:51.822 "bdev_nvme_start_mdns_discovery", 00:04:51.822 "bdev_nvme_set_multipath_policy", 00:04:51.822 "bdev_nvme_set_preferred_path", 00:04:51.822 "bdev_nvme_get_io_paths", 00:04:51.822 "bdev_nvme_remove_error_injection", 00:04:51.822 "bdev_nvme_add_error_injection", 00:04:51.822 "bdev_nvme_get_discovery_info", 00:04:51.822 "bdev_nvme_stop_discovery", 00:04:51.822 "bdev_nvme_start_discovery", 00:04:51.822 "bdev_nvme_get_controller_health_info", 00:04:51.822 "bdev_nvme_disable_controller", 00:04:51.822 "bdev_nvme_enable_controller", 00:04:51.822 "bdev_nvme_reset_controller", 00:04:51.822 "bdev_nvme_get_transport_statistics", 00:04:51.822 "bdev_nvme_apply_firmware", 00:04:51.822 "bdev_nvme_detach_controller", 00:04:51.822 "bdev_nvme_get_controllers", 00:04:51.822 "bdev_nvme_attach_controller", 00:04:51.822 "bdev_nvme_set_hotplug", 00:04:51.822 "bdev_nvme_set_options", 00:04:51.822 "bdev_passthru_delete", 00:04:51.822 "bdev_passthru_create", 00:04:51.822 "bdev_lvol_set_parent_bdev", 00:04:51.822 "bdev_lvol_set_parent", 00:04:51.822 "bdev_lvol_check_shallow_copy", 00:04:51.822 "bdev_lvol_start_shallow_copy", 00:04:51.822 "bdev_lvol_grow_lvstore", 00:04:51.822 "bdev_lvol_get_lvols", 00:04:51.822 "bdev_lvol_get_lvstores", 00:04:51.822 "bdev_lvol_delete", 00:04:51.822 "bdev_lvol_set_read_only", 00:04:51.822 "bdev_lvol_resize", 00:04:51.822 "bdev_lvol_decouple_parent", 00:04:51.822 "bdev_lvol_inflate", 00:04:51.822 "bdev_lvol_rename", 00:04:51.822 "bdev_lvol_clone_bdev", 00:04:51.822 "bdev_lvol_clone", 00:04:51.822 "bdev_lvol_snapshot", 00:04:51.822 "bdev_lvol_create", 00:04:51.822 "bdev_lvol_delete_lvstore", 00:04:51.822 "bdev_lvol_rename_lvstore", 00:04:51.822 "bdev_lvol_create_lvstore", 00:04:51.822 "bdev_raid_set_options", 00:04:51.822 "bdev_raid_remove_base_bdev", 00:04:51.822 "bdev_raid_add_base_bdev", 00:04:51.822 "bdev_raid_delete", 00:04:51.822 "bdev_raid_create", 00:04:51.822 "bdev_raid_get_bdevs", 00:04:51.822 "bdev_error_inject_error", 00:04:51.822 "bdev_error_delete", 00:04:51.822 "bdev_error_create", 00:04:51.822 "bdev_split_delete", 00:04:51.822 "bdev_split_create", 00:04:51.822 "bdev_delay_delete", 00:04:51.822 "bdev_delay_create", 00:04:51.822 "bdev_delay_update_latency", 00:04:51.822 "bdev_zone_block_delete", 00:04:51.822 "bdev_zone_block_create", 00:04:51.822 "blobfs_create", 00:04:51.822 "blobfs_detect", 00:04:51.822 "blobfs_set_cache_size", 00:04:51.822 "bdev_aio_delete", 00:04:51.822 "bdev_aio_rescan", 00:04:51.822 "bdev_aio_create", 00:04:51.822 "bdev_ftl_set_property", 00:04:51.822 "bdev_ftl_get_properties", 00:04:51.822 "bdev_ftl_get_stats", 00:04:51.822 "bdev_ftl_unmap", 00:04:51.822 "bdev_ftl_unload", 00:04:51.822 "bdev_ftl_delete", 00:04:51.822 "bdev_ftl_load", 00:04:51.822 "bdev_ftl_create", 00:04:51.822 "bdev_virtio_attach_controller", 00:04:51.822 "bdev_virtio_scsi_get_devices", 00:04:51.822 "bdev_virtio_detach_controller", 00:04:51.822 "bdev_virtio_blk_set_hotplug", 00:04:51.822 "bdev_iscsi_delete", 00:04:51.822 "bdev_iscsi_create", 00:04:51.822 "bdev_iscsi_set_options", 00:04:51.822 "accel_error_inject_error", 00:04:51.822 "ioat_scan_accel_module", 00:04:51.822 "dsa_scan_accel_module", 00:04:51.822 "iaa_scan_accel_module", 00:04:51.822 "vfu_virtio_create_scsi_endpoint", 00:04:51.822 "vfu_virtio_scsi_remove_target", 00:04:51.822 "vfu_virtio_scsi_add_target", 00:04:51.822 "vfu_virtio_create_blk_endpoint", 00:04:51.822 "vfu_virtio_delete_endpoint", 00:04:51.822 "keyring_file_remove_key", 00:04:51.822 "keyring_file_add_key", 00:04:51.822 "keyring_linux_set_options", 00:04:51.822 "iscsi_get_histogram", 00:04:51.822 "iscsi_enable_histogram", 00:04:51.822 "iscsi_set_options", 00:04:51.822 "iscsi_get_auth_groups", 00:04:51.822 "iscsi_auth_group_remove_secret", 00:04:51.822 "iscsi_auth_group_add_secret", 00:04:51.822 "iscsi_delete_auth_group", 00:04:51.822 "iscsi_create_auth_group", 00:04:51.822 "iscsi_set_discovery_auth", 00:04:51.822 "iscsi_get_options", 00:04:51.822 "iscsi_target_node_request_logout", 00:04:51.822 "iscsi_target_node_set_redirect", 00:04:51.822 "iscsi_target_node_set_auth", 00:04:51.822 "iscsi_target_node_add_lun", 00:04:51.822 "iscsi_get_stats", 00:04:51.822 "iscsi_get_connections", 00:04:51.822 "iscsi_portal_group_set_auth", 00:04:51.822 "iscsi_start_portal_group", 00:04:51.822 "iscsi_delete_portal_group", 00:04:51.822 "iscsi_create_portal_group", 00:04:51.822 "iscsi_get_portal_groups", 00:04:51.822 "iscsi_delete_target_node", 00:04:51.822 "iscsi_target_node_remove_pg_ig_maps", 00:04:51.822 "iscsi_target_node_add_pg_ig_maps", 00:04:51.822 "iscsi_create_target_node", 00:04:51.822 "iscsi_get_target_nodes", 00:04:51.822 "iscsi_delete_initiator_group", 00:04:51.822 "iscsi_initiator_group_remove_initiators", 00:04:51.822 "iscsi_initiator_group_add_initiators", 00:04:51.822 "iscsi_create_initiator_group", 00:04:51.822 "iscsi_get_initiator_groups", 00:04:51.822 "nvmf_set_crdt", 00:04:51.822 "nvmf_set_config", 00:04:51.822 "nvmf_set_max_subsystems", 00:04:51.822 "nvmf_stop_mdns_prr", 00:04:51.822 "nvmf_publish_mdns_prr", 00:04:51.822 "nvmf_subsystem_get_listeners", 00:04:51.822 "nvmf_subsystem_get_qpairs", 00:04:51.822 "nvmf_subsystem_get_controllers", 00:04:51.822 "nvmf_get_stats", 00:04:51.822 "nvmf_get_transports", 00:04:51.822 "nvmf_create_transport", 00:04:51.822 "nvmf_get_targets", 00:04:51.822 "nvmf_delete_target", 00:04:51.822 "nvmf_create_target", 00:04:51.822 "nvmf_subsystem_allow_any_host", 00:04:51.822 "nvmf_subsystem_remove_host", 00:04:51.822 "nvmf_subsystem_add_host", 00:04:51.822 "nvmf_ns_remove_host", 00:04:51.822 "nvmf_ns_add_host", 00:04:51.822 "nvmf_subsystem_remove_ns", 00:04:51.822 "nvmf_subsystem_add_ns", 00:04:51.822 "nvmf_subsystem_listener_set_ana_state", 00:04:51.822 "nvmf_discovery_get_referrals", 00:04:51.822 "nvmf_discovery_remove_referral", 00:04:51.822 "nvmf_discovery_add_referral", 00:04:51.822 "nvmf_subsystem_remove_listener", 00:04:51.822 "nvmf_subsystem_add_listener", 00:04:51.822 "nvmf_delete_subsystem", 00:04:51.822 "nvmf_create_subsystem", 00:04:51.822 "nvmf_get_subsystems", 00:04:51.822 "env_dpdk_get_mem_stats", 00:04:51.822 "nbd_get_disks", 00:04:51.822 "nbd_stop_disk", 00:04:51.822 "nbd_start_disk", 00:04:51.822 "ublk_recover_disk", 00:04:51.822 "ublk_get_disks", 00:04:51.822 "ublk_stop_disk", 00:04:51.822 "ublk_start_disk", 00:04:51.822 "ublk_destroy_target", 00:04:51.822 "ublk_create_target", 00:04:51.822 "virtio_blk_create_transport", 00:04:51.822 "virtio_blk_get_transports", 00:04:51.822 "vhost_controller_set_coalescing", 00:04:51.822 "vhost_get_controllers", 00:04:51.822 "vhost_delete_controller", 00:04:51.822 "vhost_create_blk_controller", 00:04:51.822 "vhost_scsi_controller_remove_target", 00:04:51.822 "vhost_scsi_controller_add_target", 00:04:51.822 "vhost_start_scsi_controller", 00:04:51.822 "vhost_create_scsi_controller", 00:04:51.822 "thread_set_cpumask", 00:04:51.822 "framework_get_governor", 00:04:51.822 "framework_get_scheduler", 00:04:51.822 "framework_set_scheduler", 00:04:51.822 "framework_get_reactors", 00:04:51.822 "thread_get_io_channels", 00:04:51.822 "thread_get_pollers", 00:04:51.822 "thread_get_stats", 00:04:51.822 "framework_monitor_context_switch", 00:04:51.822 "spdk_kill_instance", 00:04:51.822 "log_enable_timestamps", 00:04:51.822 "log_get_flags", 00:04:51.822 "log_clear_flag", 00:04:51.822 "log_set_flag", 00:04:51.822 "log_get_level", 00:04:51.822 "log_set_level", 00:04:51.822 "log_get_print_level", 00:04:51.822 "log_set_print_level", 00:04:51.822 "framework_enable_cpumask_locks", 00:04:51.822 "framework_disable_cpumask_locks", 00:04:51.822 "framework_wait_init", 00:04:51.822 "framework_start_init", 00:04:51.822 "scsi_get_devices", 00:04:51.822 "bdev_get_histogram", 00:04:51.822 "bdev_enable_histogram", 00:04:51.822 "bdev_set_qos_limit", 00:04:51.822 "bdev_set_qd_sampling_period", 00:04:51.822 "bdev_get_bdevs", 00:04:51.822 "bdev_reset_iostat", 00:04:51.822 "bdev_get_iostat", 00:04:51.822 "bdev_examine", 00:04:51.822 "bdev_wait_for_examine", 00:04:51.822 "bdev_set_options", 00:04:51.822 "notify_get_notifications", 00:04:51.822 "notify_get_types", 00:04:51.822 "accel_get_stats", 00:04:51.822 "accel_set_options", 00:04:51.822 "accel_set_driver", 00:04:51.822 "accel_crypto_key_destroy", 00:04:51.822 "accel_crypto_keys_get", 00:04:51.822 "accel_crypto_key_create", 00:04:51.822 "accel_assign_opc", 00:04:51.822 "accel_get_module_info", 00:04:51.822 "accel_get_opc_assignments", 00:04:51.822 "vmd_rescan", 00:04:51.822 "vmd_remove_device", 00:04:51.822 "vmd_enable", 00:04:51.822 "sock_get_default_impl", 00:04:51.822 "sock_set_default_impl", 00:04:51.822 "sock_impl_set_options", 00:04:51.822 "sock_impl_get_options", 00:04:51.822 "iobuf_get_stats", 00:04:51.822 "iobuf_set_options", 00:04:51.822 "keyring_get_keys", 00:04:51.822 "framework_get_pci_devices", 00:04:51.822 "framework_get_config", 00:04:51.822 "framework_get_subsystems", 00:04:51.822 "vfu_tgt_set_base_path", 00:04:51.822 "trace_get_info", 00:04:51.822 "trace_get_tpoint_group_mask", 00:04:51.822 "trace_disable_tpoint_group", 00:04:51.822 "trace_enable_tpoint_group", 00:04:51.822 "trace_clear_tpoint_mask", 00:04:51.822 "trace_set_tpoint_mask", 00:04:51.822 "spdk_get_version", 00:04:51.822 "rpc_get_methods" 00:04:51.822 ] 00:04:51.822 17:50:45 spdkcli_tcp -- spdkcli/tcp.sh@35 -- # timing_exit run_spdk_tgt_tcp 00:04:51.822 17:50:45 spdkcli_tcp -- common/autotest_common.sh@728 -- # xtrace_disable 00:04:51.822 17:50:45 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:04:51.822 17:50:45 spdkcli_tcp -- spdkcli/tcp.sh@37 -- # trap - SIGINT SIGTERM EXIT 00:04:51.822 17:50:45 spdkcli_tcp -- spdkcli/tcp.sh@38 -- # killprocess 410687 00:04:51.822 17:50:45 spdkcli_tcp -- common/autotest_common.sh@948 -- # '[' -z 410687 ']' 00:04:51.822 17:50:45 spdkcli_tcp -- common/autotest_common.sh@952 -- # kill -0 410687 00:04:51.822 17:50:45 spdkcli_tcp -- common/autotest_common.sh@953 -- # uname 00:04:51.822 17:50:45 spdkcli_tcp -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:04:51.822 17:50:45 spdkcli_tcp -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 410687 00:04:51.822 17:50:45 spdkcli_tcp -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:04:51.822 17:50:45 spdkcli_tcp -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:04:51.822 17:50:45 spdkcli_tcp -- common/autotest_common.sh@966 -- # echo 'killing process with pid 410687' 00:04:51.822 killing process with pid 410687 00:04:51.822 17:50:45 spdkcli_tcp -- common/autotest_common.sh@967 -- # kill 410687 00:04:51.822 17:50:45 spdkcli_tcp -- common/autotest_common.sh@972 -- # wait 410687 00:04:52.082 00:04:52.082 real 0m1.513s 00:04:52.082 user 0m2.827s 00:04:52.082 sys 0m0.422s 00:04:52.082 17:50:45 spdkcli_tcp -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:52.082 17:50:45 spdkcli_tcp -- common/autotest_common.sh@10 -- # set +x 00:04:52.082 ************************************ 00:04:52.082 END TEST spdkcli_tcp 00:04:52.082 ************************************ 00:04:52.082 17:50:45 -- common/autotest_common.sh@1142 -- # return 0 00:04:52.082 17:50:45 -- spdk/autotest.sh@180 -- # run_test dpdk_mem_utility /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:04:52.082 17:50:45 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:52.082 17:50:45 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:52.082 17:50:45 -- common/autotest_common.sh@10 -- # set +x 00:04:52.082 ************************************ 00:04:52.082 START TEST dpdk_mem_utility 00:04:52.082 ************************************ 00:04:52.082 17:50:45 dpdk_mem_utility -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/dpdk_memory_utility/test_dpdk_mem_info.sh 00:04:52.341 * Looking for test storage... 00:04:52.341 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/dpdk_memory_utility 00:04:52.341 17:50:45 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@10 -- # MEM_SCRIPT=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/dpdk_mem_info.py 00:04:52.341 17:50:45 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@13 -- # spdkpid=410991 00:04:52.341 17:50:45 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@15 -- # waitforlisten 410991 00:04:52.341 17:50:45 dpdk_mem_utility -- common/autotest_common.sh@829 -- # '[' -z 410991 ']' 00:04:52.341 17:50:45 dpdk_mem_utility -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:52.341 17:50:45 dpdk_mem_utility -- common/autotest_common.sh@834 -- # local max_retries=100 00:04:52.341 17:50:45 dpdk_mem_utility -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:52.341 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:52.341 17:50:45 dpdk_mem_utility -- common/autotest_common.sh@838 -- # xtrace_disable 00:04:52.341 17:50:45 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:04:52.341 17:50:45 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:04:52.341 [2024-07-15 17:50:45.897922] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:04:52.341 [2024-07-15 17:50:45.897973] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid410991 ] 00:04:52.341 EAL: No free 2048 kB hugepages reported on node 1 00:04:52.341 [2024-07-15 17:50:45.952316] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:52.341 [2024-07-15 17:50:46.033162] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:04:53.277 17:50:46 dpdk_mem_utility -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:04:53.277 17:50:46 dpdk_mem_utility -- common/autotest_common.sh@862 -- # return 0 00:04:53.277 17:50:46 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@17 -- # trap 'killprocess $spdkpid' SIGINT SIGTERM EXIT 00:04:53.277 17:50:46 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@19 -- # rpc_cmd env_dpdk_get_mem_stats 00:04:53.277 17:50:46 dpdk_mem_utility -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:53.277 17:50:46 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:04:53.277 { 00:04:53.277 "filename": "/tmp/spdk_mem_dump.txt" 00:04:53.277 } 00:04:53.277 17:50:46 dpdk_mem_utility -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:53.277 17:50:46 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/dpdk_mem_info.py 00:04:53.278 DPDK memory size 814.000000 MiB in 1 heap(s) 00:04:53.278 1 heaps totaling size 814.000000 MiB 00:04:53.278 size: 814.000000 MiB heap id: 0 00:04:53.278 end heaps---------- 00:04:53.278 8 mempools totaling size 598.116089 MiB 00:04:53.278 size: 212.674988 MiB name: PDU_immediate_data_Pool 00:04:53.278 size: 158.602051 MiB name: PDU_data_out_Pool 00:04:53.278 size: 84.521057 MiB name: bdev_io_410991 00:04:53.278 size: 51.011292 MiB name: evtpool_410991 00:04:53.278 size: 50.003479 MiB name: msgpool_410991 00:04:53.278 size: 21.763794 MiB name: PDU_Pool 00:04:53.278 size: 19.513306 MiB name: SCSI_TASK_Pool 00:04:53.278 size: 0.026123 MiB name: Session_Pool 00:04:53.278 end mempools------- 00:04:53.278 6 memzones totaling size 4.142822 MiB 00:04:53.278 size: 1.000366 MiB name: RG_ring_0_410991 00:04:53.278 size: 1.000366 MiB name: RG_ring_1_410991 00:04:53.278 size: 1.000366 MiB name: RG_ring_4_410991 00:04:53.278 size: 1.000366 MiB name: RG_ring_5_410991 00:04:53.278 size: 0.125366 MiB name: RG_ring_2_410991 00:04:53.278 size: 0.015991 MiB name: RG_ring_3_410991 00:04:53.278 end memzones------- 00:04:53.278 17:50:46 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/dpdk_mem_info.py -m 0 00:04:53.278 heap id: 0 total size: 814.000000 MiB number of busy elements: 41 number of free elements: 15 00:04:53.278 list of free elements. size: 12.519348 MiB 00:04:53.278 element at address: 0x200000400000 with size: 1.999512 MiB 00:04:53.278 element at address: 0x200018e00000 with size: 0.999878 MiB 00:04:53.278 element at address: 0x200019000000 with size: 0.999878 MiB 00:04:53.278 element at address: 0x200003e00000 with size: 0.996277 MiB 00:04:53.278 element at address: 0x200031c00000 with size: 0.994446 MiB 00:04:53.278 element at address: 0x200013800000 with size: 0.978699 MiB 00:04:53.278 element at address: 0x200007000000 with size: 0.959839 MiB 00:04:53.278 element at address: 0x200019200000 with size: 0.936584 MiB 00:04:53.278 element at address: 0x200000200000 with size: 0.841614 MiB 00:04:53.278 element at address: 0x20001aa00000 with size: 0.582886 MiB 00:04:53.278 element at address: 0x20000b200000 with size: 0.490723 MiB 00:04:53.278 element at address: 0x200000800000 with size: 0.487793 MiB 00:04:53.278 element at address: 0x200019400000 with size: 0.485657 MiB 00:04:53.278 element at address: 0x200027e00000 with size: 0.410034 MiB 00:04:53.278 element at address: 0x200003a00000 with size: 0.355530 MiB 00:04:53.278 list of standard malloc elements. size: 199.218079 MiB 00:04:53.278 element at address: 0x20000b3fff80 with size: 132.000122 MiB 00:04:53.278 element at address: 0x2000071fff80 with size: 64.000122 MiB 00:04:53.278 element at address: 0x200018efff80 with size: 1.000122 MiB 00:04:53.278 element at address: 0x2000190fff80 with size: 1.000122 MiB 00:04:53.278 element at address: 0x2000192fff80 with size: 1.000122 MiB 00:04:53.278 element at address: 0x2000003d9f00 with size: 0.140747 MiB 00:04:53.278 element at address: 0x2000192eff00 with size: 0.062622 MiB 00:04:53.278 element at address: 0x2000003fdf80 with size: 0.007935 MiB 00:04:53.278 element at address: 0x2000192efdc0 with size: 0.000305 MiB 00:04:53.278 element at address: 0x2000002d7740 with size: 0.000183 MiB 00:04:53.278 element at address: 0x2000002d7800 with size: 0.000183 MiB 00:04:53.278 element at address: 0x2000002d78c0 with size: 0.000183 MiB 00:04:53.278 element at address: 0x2000002d7ac0 with size: 0.000183 MiB 00:04:53.278 element at address: 0x2000002d7b80 with size: 0.000183 MiB 00:04:53.278 element at address: 0x2000002d7c40 with size: 0.000183 MiB 00:04:53.278 element at address: 0x2000003d9e40 with size: 0.000183 MiB 00:04:53.278 element at address: 0x20000087ce00 with size: 0.000183 MiB 00:04:53.278 element at address: 0x20000087cec0 with size: 0.000183 MiB 00:04:53.278 element at address: 0x2000008fd180 with size: 0.000183 MiB 00:04:53.278 element at address: 0x200003a5b040 with size: 0.000183 MiB 00:04:53.278 element at address: 0x200003adb300 with size: 0.000183 MiB 00:04:53.278 element at address: 0x200003adb500 with size: 0.000183 MiB 00:04:53.278 element at address: 0x200003adf7c0 with size: 0.000183 MiB 00:04:53.278 element at address: 0x200003affa80 with size: 0.000183 MiB 00:04:53.278 element at address: 0x200003affb40 with size: 0.000183 MiB 00:04:53.278 element at address: 0x200003eff0c0 with size: 0.000183 MiB 00:04:53.278 element at address: 0x2000070fdd80 with size: 0.000183 MiB 00:04:53.278 element at address: 0x20000b27da00 with size: 0.000183 MiB 00:04:53.278 element at address: 0x20000b27dac0 with size: 0.000183 MiB 00:04:53.278 element at address: 0x20000b2fdd80 with size: 0.000183 MiB 00:04:53.278 element at address: 0x2000138fa8c0 with size: 0.000183 MiB 00:04:53.278 element at address: 0x2000192efc40 with size: 0.000183 MiB 00:04:53.278 element at address: 0x2000192efd00 with size: 0.000183 MiB 00:04:53.278 element at address: 0x2000194bc740 with size: 0.000183 MiB 00:04:53.278 element at address: 0x20001aa95380 with size: 0.000183 MiB 00:04:53.278 element at address: 0x20001aa95440 with size: 0.000183 MiB 00:04:53.278 element at address: 0x200027e68f80 with size: 0.000183 MiB 00:04:53.278 element at address: 0x200027e69040 with size: 0.000183 MiB 00:04:53.278 element at address: 0x200027e6fc40 with size: 0.000183 MiB 00:04:53.278 element at address: 0x200027e6fe40 with size: 0.000183 MiB 00:04:53.278 element at address: 0x200027e6ff00 with size: 0.000183 MiB 00:04:53.278 list of memzone associated elements. size: 602.262573 MiB 00:04:53.278 element at address: 0x20001aa95500 with size: 211.416748 MiB 00:04:53.278 associated memzone info: size: 211.416626 MiB name: MP_PDU_immediate_data_Pool_0 00:04:53.278 element at address: 0x200027e6ffc0 with size: 157.562561 MiB 00:04:53.278 associated memzone info: size: 157.562439 MiB name: MP_PDU_data_out_Pool_0 00:04:53.278 element at address: 0x2000139fab80 with size: 84.020630 MiB 00:04:53.278 associated memzone info: size: 84.020508 MiB name: MP_bdev_io_410991_0 00:04:53.278 element at address: 0x2000009ff380 with size: 48.003052 MiB 00:04:53.278 associated memzone info: size: 48.002930 MiB name: MP_evtpool_410991_0 00:04:53.278 element at address: 0x200003fff380 with size: 48.003052 MiB 00:04:53.278 associated memzone info: size: 48.002930 MiB name: MP_msgpool_410991_0 00:04:53.278 element at address: 0x2000195be940 with size: 20.255554 MiB 00:04:53.278 associated memzone info: size: 20.255432 MiB name: MP_PDU_Pool_0 00:04:53.278 element at address: 0x200031dfeb40 with size: 18.005066 MiB 00:04:53.278 associated memzone info: size: 18.004944 MiB name: MP_SCSI_TASK_Pool_0 00:04:53.278 element at address: 0x2000005ffe00 with size: 2.000488 MiB 00:04:53.278 associated memzone info: size: 2.000366 MiB name: RG_MP_evtpool_410991 00:04:53.278 element at address: 0x200003bffe00 with size: 2.000488 MiB 00:04:53.278 associated memzone info: size: 2.000366 MiB name: RG_MP_msgpool_410991 00:04:53.278 element at address: 0x2000002d7d00 with size: 1.008118 MiB 00:04:53.278 associated memzone info: size: 1.007996 MiB name: MP_evtpool_410991 00:04:53.278 element at address: 0x20000b2fde40 with size: 1.008118 MiB 00:04:53.278 associated memzone info: size: 1.007996 MiB name: MP_PDU_Pool 00:04:53.278 element at address: 0x2000194bc800 with size: 1.008118 MiB 00:04:53.278 associated memzone info: size: 1.007996 MiB name: MP_PDU_immediate_data_Pool 00:04:53.278 element at address: 0x2000070fde40 with size: 1.008118 MiB 00:04:53.278 associated memzone info: size: 1.007996 MiB name: MP_PDU_data_out_Pool 00:04:53.278 element at address: 0x2000008fd240 with size: 1.008118 MiB 00:04:53.278 associated memzone info: size: 1.007996 MiB name: MP_SCSI_TASK_Pool 00:04:53.278 element at address: 0x200003eff180 with size: 1.000488 MiB 00:04:53.278 associated memzone info: size: 1.000366 MiB name: RG_ring_0_410991 00:04:53.278 element at address: 0x200003affc00 with size: 1.000488 MiB 00:04:53.278 associated memzone info: size: 1.000366 MiB name: RG_ring_1_410991 00:04:53.278 element at address: 0x2000138fa980 with size: 1.000488 MiB 00:04:53.278 associated memzone info: size: 1.000366 MiB name: RG_ring_4_410991 00:04:53.278 element at address: 0x200031cfe940 with size: 1.000488 MiB 00:04:53.278 associated memzone info: size: 1.000366 MiB name: RG_ring_5_410991 00:04:53.278 element at address: 0x200003a5b100 with size: 0.500488 MiB 00:04:53.278 associated memzone info: size: 0.500366 MiB name: RG_MP_bdev_io_410991 00:04:53.278 element at address: 0x20000b27db80 with size: 0.500488 MiB 00:04:53.278 associated memzone info: size: 0.500366 MiB name: RG_MP_PDU_Pool 00:04:53.278 element at address: 0x20000087cf80 with size: 0.500488 MiB 00:04:53.278 associated memzone info: size: 0.500366 MiB name: RG_MP_SCSI_TASK_Pool 00:04:53.278 element at address: 0x20001947c540 with size: 0.250488 MiB 00:04:53.278 associated memzone info: size: 0.250366 MiB name: RG_MP_PDU_immediate_data_Pool 00:04:53.278 element at address: 0x200003adf880 with size: 0.125488 MiB 00:04:53.278 associated memzone info: size: 0.125366 MiB name: RG_ring_2_410991 00:04:53.278 element at address: 0x2000070f5b80 with size: 0.031738 MiB 00:04:53.278 associated memzone info: size: 0.031616 MiB name: RG_MP_PDU_data_out_Pool 00:04:53.278 element at address: 0x200027e69100 with size: 0.023743 MiB 00:04:53.278 associated memzone info: size: 0.023621 MiB name: MP_Session_Pool_0 00:04:53.278 element at address: 0x200003adb5c0 with size: 0.016113 MiB 00:04:53.278 associated memzone info: size: 0.015991 MiB name: RG_ring_3_410991 00:04:53.278 element at address: 0x200027e6f240 with size: 0.002441 MiB 00:04:53.278 associated memzone info: size: 0.002319 MiB name: RG_MP_Session_Pool 00:04:53.278 element at address: 0x2000002d7980 with size: 0.000305 MiB 00:04:53.278 associated memzone info: size: 0.000183 MiB name: MP_msgpool_410991 00:04:53.278 element at address: 0x200003adb3c0 with size: 0.000305 MiB 00:04:53.278 associated memzone info: size: 0.000183 MiB name: MP_bdev_io_410991 00:04:53.278 element at address: 0x200027e6fd00 with size: 0.000305 MiB 00:04:53.278 associated memzone info: size: 0.000183 MiB name: MP_Session_Pool 00:04:53.278 17:50:46 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@25 -- # trap - SIGINT SIGTERM EXIT 00:04:53.278 17:50:46 dpdk_mem_utility -- dpdk_memory_utility/test_dpdk_mem_info.sh@26 -- # killprocess 410991 00:04:53.278 17:50:46 dpdk_mem_utility -- common/autotest_common.sh@948 -- # '[' -z 410991 ']' 00:04:53.278 17:50:46 dpdk_mem_utility -- common/autotest_common.sh@952 -- # kill -0 410991 00:04:53.279 17:50:46 dpdk_mem_utility -- common/autotest_common.sh@953 -- # uname 00:04:53.279 17:50:46 dpdk_mem_utility -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:04:53.279 17:50:46 dpdk_mem_utility -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 410991 00:04:53.279 17:50:46 dpdk_mem_utility -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:04:53.279 17:50:46 dpdk_mem_utility -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:04:53.279 17:50:46 dpdk_mem_utility -- common/autotest_common.sh@966 -- # echo 'killing process with pid 410991' 00:04:53.279 killing process with pid 410991 00:04:53.279 17:50:46 dpdk_mem_utility -- common/autotest_common.sh@967 -- # kill 410991 00:04:53.279 17:50:46 dpdk_mem_utility -- common/autotest_common.sh@972 -- # wait 410991 00:04:53.537 00:04:53.537 real 0m1.356s 00:04:53.537 user 0m1.432s 00:04:53.537 sys 0m0.373s 00:04:53.537 17:50:47 dpdk_mem_utility -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:53.537 17:50:47 dpdk_mem_utility -- common/autotest_common.sh@10 -- # set +x 00:04:53.537 ************************************ 00:04:53.537 END TEST dpdk_mem_utility 00:04:53.537 ************************************ 00:04:53.538 17:50:47 -- common/autotest_common.sh@1142 -- # return 0 00:04:53.538 17:50:47 -- spdk/autotest.sh@181 -- # run_test event /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event.sh 00:04:53.538 17:50:47 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:53.538 17:50:47 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:53.538 17:50:47 -- common/autotest_common.sh@10 -- # set +x 00:04:53.538 ************************************ 00:04:53.538 START TEST event 00:04:53.538 ************************************ 00:04:53.538 17:50:47 event -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event.sh 00:04:53.795 * Looking for test storage... 00:04:53.795 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event 00:04:53.795 17:50:47 event -- event/event.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/nbd_common.sh 00:04:53.795 17:50:47 event -- bdev/nbd_common.sh@6 -- # set -e 00:04:53.795 17:50:47 event -- event/event.sh@45 -- # run_test event_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:04:53.796 17:50:47 event -- common/autotest_common.sh@1099 -- # '[' 6 -le 1 ']' 00:04:53.796 17:50:47 event -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:53.796 17:50:47 event -- common/autotest_common.sh@10 -- # set +x 00:04:53.796 ************************************ 00:04:53.796 START TEST event_perf 00:04:53.796 ************************************ 00:04:53.796 17:50:47 event.event_perf -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/event_perf/event_perf -m 0xF -t 1 00:04:53.796 Running I/O for 1 seconds...[2024-07-15 17:50:47.325444] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:04:53.796 [2024-07-15 17:50:47.325487] [ DPDK EAL parameters: event_perf --no-shconf -c 0xF --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid411285 ] 00:04:53.796 EAL: No free 2048 kB hugepages reported on node 1 00:04:53.796 [2024-07-15 17:50:47.379854] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:04:53.796 [2024-07-15 17:50:47.456154] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:04:53.796 [2024-07-15 17:50:47.456175] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:04:53.796 [2024-07-15 17:50:47.456268] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:04:53.796 [2024-07-15 17:50:47.456270] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:04:55.173 Running I/O for 1 seconds... 00:04:55.173 lcore 0: 204874 00:04:55.173 lcore 1: 204872 00:04:55.173 lcore 2: 204874 00:04:55.173 lcore 3: 204874 00:04:55.173 done. 00:04:55.173 00:04:55.173 real 0m1.214s 00:04:55.173 user 0m4.137s 00:04:55.173 sys 0m0.073s 00:04:55.173 17:50:48 event.event_perf -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:55.173 17:50:48 event.event_perf -- common/autotest_common.sh@10 -- # set +x 00:04:55.173 ************************************ 00:04:55.173 END TEST event_perf 00:04:55.173 ************************************ 00:04:55.173 17:50:48 event -- common/autotest_common.sh@1142 -- # return 0 00:04:55.173 17:50:48 event -- event/event.sh@46 -- # run_test event_reactor /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor/reactor -t 1 00:04:55.173 17:50:48 event -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:04:55.173 17:50:48 event -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:55.173 17:50:48 event -- common/autotest_common.sh@10 -- # set +x 00:04:55.173 ************************************ 00:04:55.173 START TEST event_reactor 00:04:55.173 ************************************ 00:04:55.173 17:50:48 event.event_reactor -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor/reactor -t 1 00:04:55.173 [2024-07-15 17:50:48.618774] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:04:55.173 [2024-07-15 17:50:48.618848] [ DPDK EAL parameters: reactor --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid411537 ] 00:04:55.173 EAL: No free 2048 kB hugepages reported on node 1 00:04:55.173 [2024-07-15 17:50:48.676386] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:55.173 [2024-07-15 17:50:48.747494] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:04:56.110 test_start 00:04:56.110 oneshot 00:04:56.110 tick 100 00:04:56.110 tick 100 00:04:56.110 tick 250 00:04:56.110 tick 100 00:04:56.110 tick 100 00:04:56.110 tick 250 00:04:56.110 tick 100 00:04:56.110 tick 500 00:04:56.110 tick 100 00:04:56.110 tick 100 00:04:56.110 tick 250 00:04:56.110 tick 100 00:04:56.110 tick 100 00:04:56.110 test_end 00:04:56.110 00:04:56.110 real 0m1.217s 00:04:56.110 user 0m1.134s 00:04:56.110 sys 0m0.078s 00:04:56.110 17:50:49 event.event_reactor -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:56.110 17:50:49 event.event_reactor -- common/autotest_common.sh@10 -- # set +x 00:04:56.110 ************************************ 00:04:56.110 END TEST event_reactor 00:04:56.110 ************************************ 00:04:56.382 17:50:49 event -- common/autotest_common.sh@1142 -- # return 0 00:04:56.382 17:50:49 event -- event/event.sh@47 -- # run_test event_reactor_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor_perf/reactor_perf -t 1 00:04:56.382 17:50:49 event -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:04:56.382 17:50:49 event -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:56.382 17:50:49 event -- common/autotest_common.sh@10 -- # set +x 00:04:56.382 ************************************ 00:04:56.382 START TEST event_reactor_perf 00:04:56.383 ************************************ 00:04:56.383 17:50:49 event.event_reactor_perf -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/reactor_perf/reactor_perf -t 1 00:04:56.383 [2024-07-15 17:50:49.898784] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:04:56.383 [2024-07-15 17:50:49.898850] [ DPDK EAL parameters: reactor_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid411784 ] 00:04:56.383 EAL: No free 2048 kB hugepages reported on node 1 00:04:56.383 [2024-07-15 17:50:49.956706] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:04:56.383 [2024-07-15 17:50:50.036583] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:04:57.762 test_start 00:04:57.762 test_end 00:04:57.762 Performance: 483661 events per second 00:04:57.762 00:04:57.763 real 0m1.229s 00:04:57.763 user 0m1.153s 00:04:57.763 sys 0m0.071s 00:04:57.763 17:50:51 event.event_reactor_perf -- common/autotest_common.sh@1124 -- # xtrace_disable 00:04:57.763 17:50:51 event.event_reactor_perf -- common/autotest_common.sh@10 -- # set +x 00:04:57.763 ************************************ 00:04:57.763 END TEST event_reactor_perf 00:04:57.763 ************************************ 00:04:57.763 17:50:51 event -- common/autotest_common.sh@1142 -- # return 0 00:04:57.763 17:50:51 event -- event/event.sh@49 -- # uname -s 00:04:57.763 17:50:51 event -- event/event.sh@49 -- # '[' Linux = Linux ']' 00:04:57.763 17:50:51 event -- event/event.sh@50 -- # run_test event_scheduler /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler/scheduler.sh 00:04:57.763 17:50:51 event -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:57.763 17:50:51 event -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:57.763 17:50:51 event -- common/autotest_common.sh@10 -- # set +x 00:04:57.763 ************************************ 00:04:57.763 START TEST event_scheduler 00:04:57.763 ************************************ 00:04:57.763 17:50:51 event.event_scheduler -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler/scheduler.sh 00:04:57.763 * Looking for test storage... 00:04:57.763 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler 00:04:57.763 17:50:51 event.event_scheduler -- scheduler/scheduler.sh@29 -- # rpc=rpc_cmd 00:04:57.763 17:50:51 event.event_scheduler -- scheduler/scheduler.sh@35 -- # scheduler_pid=412062 00:04:57.763 17:50:51 event.event_scheduler -- scheduler/scheduler.sh@36 -- # trap 'killprocess $scheduler_pid; exit 1' SIGINT SIGTERM EXIT 00:04:57.763 17:50:51 event.event_scheduler -- scheduler/scheduler.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/scheduler/scheduler -m 0xF -p 0x2 --wait-for-rpc -f 00:04:57.763 17:50:51 event.event_scheduler -- scheduler/scheduler.sh@37 -- # waitforlisten 412062 00:04:57.763 17:50:51 event.event_scheduler -- common/autotest_common.sh@829 -- # '[' -z 412062 ']' 00:04:57.763 17:50:51 event.event_scheduler -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:04:57.763 17:50:51 event.event_scheduler -- common/autotest_common.sh@834 -- # local max_retries=100 00:04:57.763 17:50:51 event.event_scheduler -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:04:57.763 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:04:57.763 17:50:51 event.event_scheduler -- common/autotest_common.sh@838 -- # xtrace_disable 00:04:57.763 17:50:51 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:04:57.763 [2024-07-15 17:50:51.301825] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:04:57.763 [2024-07-15 17:50:51.301873] [ DPDK EAL parameters: scheduler --no-shconf -c 0xF --main-lcore=2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid412062 ] 00:04:57.763 EAL: No free 2048 kB hugepages reported on node 1 00:04:57.763 [2024-07-15 17:50:51.352134] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:04:57.763 [2024-07-15 17:50:51.428118] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:04:57.763 [2024-07-15 17:50:51.428203] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:04:57.763 [2024-07-15 17:50:51.428293] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:04:57.763 [2024-07-15 17:50:51.428296] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:04:58.701 17:50:52 event.event_scheduler -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:04:58.701 17:50:52 event.event_scheduler -- common/autotest_common.sh@862 -- # return 0 00:04:58.701 17:50:52 event.event_scheduler -- scheduler/scheduler.sh@39 -- # rpc_cmd framework_set_scheduler dynamic 00:04:58.701 17:50:52 event.event_scheduler -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:58.701 17:50:52 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:04:58.701 [2024-07-15 17:50:52.126696] dpdk_governor.c: 173:_init: *ERROR*: App core mask contains some but not all of a set of SMT siblings 00:04:58.701 [2024-07-15 17:50:52.126715] scheduler_dynamic.c: 270:init: *NOTICE*: Unable to initialize dpdk governor 00:04:58.701 [2024-07-15 17:50:52.126725] scheduler_dynamic.c: 416:set_opts: *NOTICE*: Setting scheduler load limit to 20 00:04:58.701 [2024-07-15 17:50:52.126730] scheduler_dynamic.c: 418:set_opts: *NOTICE*: Setting scheduler core limit to 80 00:04:58.701 [2024-07-15 17:50:52.126735] scheduler_dynamic.c: 420:set_opts: *NOTICE*: Setting scheduler core busy to 95 00:04:58.701 17:50:52 event.event_scheduler -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:58.701 17:50:52 event.event_scheduler -- scheduler/scheduler.sh@40 -- # rpc_cmd framework_start_init 00:04:58.701 17:50:52 event.event_scheduler -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:58.701 17:50:52 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:04:58.701 [2024-07-15 17:50:52.198814] scheduler.c: 382:test_start: *NOTICE*: Scheduler test application started. 00:04:58.701 17:50:52 event.event_scheduler -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:58.701 17:50:52 event.event_scheduler -- scheduler/scheduler.sh@43 -- # run_test scheduler_create_thread scheduler_create_thread 00:04:58.701 17:50:52 event.event_scheduler -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:04:58.701 17:50:52 event.event_scheduler -- common/autotest_common.sh@1105 -- # xtrace_disable 00:04:58.701 17:50:52 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:04:58.701 ************************************ 00:04:58.701 START TEST scheduler_create_thread 00:04:58.701 ************************************ 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1123 -- # scheduler_create_thread 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@12 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x1 -a 100 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:58.701 2 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@13 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x2 -a 100 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:58.701 3 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@14 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x4 -a 100 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:58.701 4 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@15 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n active_pinned -m 0x8 -a 100 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:58.701 5 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@16 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x1 -a 0 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:58.701 6 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@17 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x2 -a 0 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:58.701 7 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@18 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x4 -a 0 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:58.701 8 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@19 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n idle_pinned -m 0x8 -a 0 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:58.701 9 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@21 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n one_third_active -a 30 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:58.701 10 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n half_active -a 0 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@22 -- # thread_id=11 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@23 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_set_active 11 50 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:58.701 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:04:59.270 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:04:59.270 17:50:52 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_create -n deleted -a 100 00:04:59.270 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:04:59.270 17:50:52 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:00.649 17:50:54 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:05:00.649 17:50:54 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@25 -- # thread_id=12 00:05:00.649 17:50:54 event.event_scheduler.scheduler_create_thread -- scheduler/scheduler.sh@26 -- # rpc_cmd --plugin scheduler_plugin scheduler_thread_delete 12 00:05:00.649 17:50:54 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@559 -- # xtrace_disable 00:05:00.649 17:50:54 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:02.029 17:50:55 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:05:02.029 00:05:02.029 real 0m3.099s 00:05:02.029 user 0m0.025s 00:05:02.029 sys 0m0.003s 00:05:02.029 17:50:55 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:02.029 17:50:55 event.event_scheduler.scheduler_create_thread -- common/autotest_common.sh@10 -- # set +x 00:05:02.029 ************************************ 00:05:02.029 END TEST scheduler_create_thread 00:05:02.029 ************************************ 00:05:02.029 17:50:55 event.event_scheduler -- common/autotest_common.sh@1142 -- # return 0 00:05:02.029 17:50:55 event.event_scheduler -- scheduler/scheduler.sh@45 -- # trap - SIGINT SIGTERM EXIT 00:05:02.029 17:50:55 event.event_scheduler -- scheduler/scheduler.sh@46 -- # killprocess 412062 00:05:02.029 17:50:55 event.event_scheduler -- common/autotest_common.sh@948 -- # '[' -z 412062 ']' 00:05:02.029 17:50:55 event.event_scheduler -- common/autotest_common.sh@952 -- # kill -0 412062 00:05:02.029 17:50:55 event.event_scheduler -- common/autotest_common.sh@953 -- # uname 00:05:02.029 17:50:55 event.event_scheduler -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:05:02.029 17:50:55 event.event_scheduler -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 412062 00:05:02.029 17:50:55 event.event_scheduler -- common/autotest_common.sh@954 -- # process_name=reactor_2 00:05:02.029 17:50:55 event.event_scheduler -- common/autotest_common.sh@958 -- # '[' reactor_2 = sudo ']' 00:05:02.029 17:50:55 event.event_scheduler -- common/autotest_common.sh@966 -- # echo 'killing process with pid 412062' 00:05:02.029 killing process with pid 412062 00:05:02.029 17:50:55 event.event_scheduler -- common/autotest_common.sh@967 -- # kill 412062 00:05:02.029 17:50:55 event.event_scheduler -- common/autotest_common.sh@972 -- # wait 412062 00:05:02.029 [2024-07-15 17:50:55.714039] scheduler.c: 360:test_shutdown: *NOTICE*: Scheduler test application stopped. 00:05:02.288 00:05:02.288 real 0m4.750s 00:05:02.288 user 0m9.317s 00:05:02.288 sys 0m0.358s 00:05:02.288 17:50:55 event.event_scheduler -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:02.288 17:50:55 event.event_scheduler -- common/autotest_common.sh@10 -- # set +x 00:05:02.288 ************************************ 00:05:02.288 END TEST event_scheduler 00:05:02.288 ************************************ 00:05:02.288 17:50:55 event -- common/autotest_common.sh@1142 -- # return 0 00:05:02.288 17:50:55 event -- event/event.sh@51 -- # modprobe -n nbd 00:05:02.288 17:50:55 event -- event/event.sh@52 -- # run_test app_repeat app_repeat_test 00:05:02.288 17:50:55 event -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:05:02.288 17:50:55 event -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:02.288 17:50:55 event -- common/autotest_common.sh@10 -- # set +x 00:05:02.288 ************************************ 00:05:02.288 START TEST app_repeat 00:05:02.288 ************************************ 00:05:02.288 17:50:55 event.app_repeat -- common/autotest_common.sh@1123 -- # app_repeat_test 00:05:02.288 17:50:55 event.app_repeat -- event/event.sh@12 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:02.288 17:50:55 event.app_repeat -- event/event.sh@13 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:02.288 17:50:55 event.app_repeat -- event/event.sh@13 -- # local nbd_list 00:05:02.288 17:50:55 event.app_repeat -- event/event.sh@14 -- # bdev_list=('Malloc0' 'Malloc1') 00:05:02.288 17:50:55 event.app_repeat -- event/event.sh@14 -- # local bdev_list 00:05:02.288 17:50:55 event.app_repeat -- event/event.sh@15 -- # local repeat_times=4 00:05:02.288 17:50:55 event.app_repeat -- event/event.sh@17 -- # modprobe nbd 00:05:02.288 17:50:55 event.app_repeat -- event/event.sh@19 -- # repeat_pid=412847 00:05:02.288 17:50:55 event.app_repeat -- event/event.sh@20 -- # trap 'killprocess $repeat_pid; exit 1' SIGINT SIGTERM EXIT 00:05:02.288 17:50:55 event.app_repeat -- event/event.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/app_repeat/app_repeat -r /var/tmp/spdk-nbd.sock -m 0x3 -t 4 00:05:02.288 17:50:55 event.app_repeat -- event/event.sh@21 -- # echo 'Process app_repeat pid: 412847' 00:05:02.288 Process app_repeat pid: 412847 00:05:02.288 17:50:55 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:05:02.288 17:50:55 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 0' 00:05:02.288 spdk_app_start Round 0 00:05:02.288 17:50:55 event.app_repeat -- event/event.sh@25 -- # waitforlisten 412847 /var/tmp/spdk-nbd.sock 00:05:02.288 17:50:55 event.app_repeat -- common/autotest_common.sh@829 -- # '[' -z 412847 ']' 00:05:02.288 17:50:55 event.app_repeat -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:05:02.288 17:50:55 event.app_repeat -- common/autotest_common.sh@834 -- # local max_retries=100 00:05:02.288 17:50:55 event.app_repeat -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:05:02.288 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:05:02.288 17:50:55 event.app_repeat -- common/autotest_common.sh@838 -- # xtrace_disable 00:05:02.288 17:50:55 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:05:02.288 [2024-07-15 17:50:56.014466] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:02.288 [2024-07-15 17:50:56.014516] [ DPDK EAL parameters: app_repeat --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid412847 ] 00:05:02.548 EAL: No free 2048 kB hugepages reported on node 1 00:05:02.548 [2024-07-15 17:50:56.070028] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:05:02.548 [2024-07-15 17:50:56.150043] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:05:02.548 [2024-07-15 17:50:56.150046] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:02.548 17:50:56 event.app_repeat -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:05:02.548 17:50:56 event.app_repeat -- common/autotest_common.sh@862 -- # return 0 00:05:02.548 17:50:56 event.app_repeat -- event/event.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:05:02.807 Malloc0 00:05:02.807 17:50:56 event.app_repeat -- event/event.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:05:03.066 Malloc1 00:05:03.067 17:50:56 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:05:03.067 17:50:56 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:03.067 17:50:56 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:05:03.067 17:50:56 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:05:03.067 17:50:56 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:03.067 17:50:56 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:05:03.067 17:50:56 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:05:03.067 17:50:56 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:03.067 17:50:56 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:05:03.067 17:50:56 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:05:03.067 17:50:56 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:03.067 17:50:56 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:05:03.067 17:50:56 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:05:03.067 17:50:56 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:05:03.067 17:50:56 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:05:03.067 17:50:56 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:05:03.067 /dev/nbd0 00:05:03.067 17:50:56 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:05:03.067 17:50:56 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:05:03.067 17:50:56 event.app_repeat -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:05:03.067 17:50:56 event.app_repeat -- common/autotest_common.sh@867 -- # local i 00:05:03.067 17:50:56 event.app_repeat -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:05:03.067 17:50:56 event.app_repeat -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:05:03.067 17:50:56 event.app_repeat -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:05:03.067 17:50:56 event.app_repeat -- common/autotest_common.sh@871 -- # break 00:05:03.067 17:50:56 event.app_repeat -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:05:03.067 17:50:56 event.app_repeat -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:05:03.067 17:50:56 event.app_repeat -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:05:03.067 1+0 records in 00:05:03.067 1+0 records out 00:05:03.067 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000178765 s, 22.9 MB/s 00:05:03.326 17:50:56 event.app_repeat -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:05:03.326 17:50:56 event.app_repeat -- common/autotest_common.sh@884 -- # size=4096 00:05:03.326 17:50:56 event.app_repeat -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:05:03.326 17:50:56 event.app_repeat -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:05:03.326 17:50:56 event.app_repeat -- common/autotest_common.sh@887 -- # return 0 00:05:03.326 17:50:56 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:05:03.326 17:50:56 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:05:03.326 17:50:56 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:05:03.326 /dev/nbd1 00:05:03.326 17:50:56 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:05:03.326 17:50:56 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:05:03.326 17:50:56 event.app_repeat -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:05:03.326 17:50:56 event.app_repeat -- common/autotest_common.sh@867 -- # local i 00:05:03.326 17:50:56 event.app_repeat -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:05:03.326 17:50:56 event.app_repeat -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:05:03.326 17:50:56 event.app_repeat -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:05:03.326 17:50:56 event.app_repeat -- common/autotest_common.sh@871 -- # break 00:05:03.326 17:50:56 event.app_repeat -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:05:03.326 17:50:56 event.app_repeat -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:05:03.326 17:50:56 event.app_repeat -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:05:03.326 1+0 records in 00:05:03.326 1+0 records out 00:05:03.326 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000160336 s, 25.5 MB/s 00:05:03.326 17:50:57 event.app_repeat -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:05:03.326 17:50:57 event.app_repeat -- common/autotest_common.sh@884 -- # size=4096 00:05:03.326 17:50:57 event.app_repeat -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:05:03.326 17:50:57 event.app_repeat -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:05:03.326 17:50:57 event.app_repeat -- common/autotest_common.sh@887 -- # return 0 00:05:03.326 17:50:57 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:05:03.326 17:50:57 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:05:03.326 17:50:57 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:05:03.326 17:50:57 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:03.326 17:50:57 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:05:03.585 17:50:57 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:05:03.585 { 00:05:03.585 "nbd_device": "/dev/nbd0", 00:05:03.585 "bdev_name": "Malloc0" 00:05:03.585 }, 00:05:03.585 { 00:05:03.585 "nbd_device": "/dev/nbd1", 00:05:03.585 "bdev_name": "Malloc1" 00:05:03.585 } 00:05:03.585 ]' 00:05:03.585 17:50:57 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:05:03.585 { 00:05:03.585 "nbd_device": "/dev/nbd0", 00:05:03.585 "bdev_name": "Malloc0" 00:05:03.585 }, 00:05:03.585 { 00:05:03.585 "nbd_device": "/dev/nbd1", 00:05:03.585 "bdev_name": "Malloc1" 00:05:03.585 } 00:05:03.585 ]' 00:05:03.585 17:50:57 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:05:03.585 17:50:57 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:05:03.585 /dev/nbd1' 00:05:03.585 17:50:57 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:05:03.585 /dev/nbd1' 00:05:03.585 17:50:57 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:05:03.585 17:50:57 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:05:03.585 17:50:57 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:05:03.585 17:50:57 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:05:03.585 17:50:57 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:05:03.585 17:50:57 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:05:03.585 17:50:57 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:03.585 17:50:57 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:05:03.585 17:50:57 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:05:03.586 256+0 records in 00:05:03.586 256+0 records out 00:05:03.586 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0103469 s, 101 MB/s 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:05:03.586 256+0 records in 00:05:03.586 256+0 records out 00:05:03.586 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0134179 s, 78.1 MB/s 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:05:03.586 256+0 records in 00:05:03.586 256+0 records out 00:05:03.586 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0146691 s, 71.5 MB/s 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:05:03.586 17:50:57 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:05:03.845 17:50:57 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:05:03.845 17:50:57 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:05:03.845 17:50:57 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:05:03.845 17:50:57 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:05:03.845 17:50:57 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:05:03.845 17:50:57 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:05:03.845 17:50:57 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:05:03.845 17:50:57 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:05:03.845 17:50:57 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:05:03.845 17:50:57 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:05:04.105 17:50:57 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:05:04.105 17:50:57 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:05:04.105 17:50:57 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:05:04.105 17:50:57 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:05:04.105 17:50:57 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:05:04.105 17:50:57 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:05:04.105 17:50:57 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:05:04.105 17:50:57 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:05:04.105 17:50:57 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:05:04.105 17:50:57 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:04.105 17:50:57 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:05:04.392 17:50:57 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:05:04.392 17:50:57 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:05:04.392 17:50:57 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:05:04.392 17:50:57 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:05:04.392 17:50:57 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:05:04.393 17:50:57 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:05:04.393 17:50:57 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:05:04.393 17:50:57 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:05:04.393 17:50:57 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:05:04.393 17:50:57 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:05:04.393 17:50:57 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:05:04.393 17:50:57 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:05:04.393 17:50:57 event.app_repeat -- event/event.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:05:04.652 17:50:58 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:05:04.652 [2024-07-15 17:50:58.319717] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:05:04.911 [2024-07-15 17:50:58.387345] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:05:04.911 [2024-07-15 17:50:58.387348] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:04.911 [2024-07-15 17:50:58.427840] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:05:04.911 [2024-07-15 17:50:58.427877] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:05:07.444 17:51:01 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:05:07.444 17:51:01 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 1' 00:05:07.444 spdk_app_start Round 1 00:05:07.444 17:51:01 event.app_repeat -- event/event.sh@25 -- # waitforlisten 412847 /var/tmp/spdk-nbd.sock 00:05:07.444 17:51:01 event.app_repeat -- common/autotest_common.sh@829 -- # '[' -z 412847 ']' 00:05:07.444 17:51:01 event.app_repeat -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:05:07.444 17:51:01 event.app_repeat -- common/autotest_common.sh@834 -- # local max_retries=100 00:05:07.444 17:51:01 event.app_repeat -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:05:07.444 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:05:07.444 17:51:01 event.app_repeat -- common/autotest_common.sh@838 -- # xtrace_disable 00:05:07.444 17:51:01 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:05:07.702 17:51:01 event.app_repeat -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:05:07.702 17:51:01 event.app_repeat -- common/autotest_common.sh@862 -- # return 0 00:05:07.702 17:51:01 event.app_repeat -- event/event.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:05:07.961 Malloc0 00:05:07.961 17:51:01 event.app_repeat -- event/event.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:05:07.961 Malloc1 00:05:07.961 17:51:01 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:05:07.961 17:51:01 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:07.961 17:51:01 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:05:07.961 17:51:01 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:05:07.961 17:51:01 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:07.961 17:51:01 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:05:07.961 17:51:01 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:05:07.961 17:51:01 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:07.961 17:51:01 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:05:07.961 17:51:01 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:05:07.961 17:51:01 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:07.961 17:51:01 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:05:07.961 17:51:01 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:05:07.961 17:51:01 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:05:07.961 17:51:01 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:05:07.961 17:51:01 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:05:08.220 /dev/nbd0 00:05:08.220 17:51:01 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:05:08.220 17:51:01 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:05:08.220 17:51:01 event.app_repeat -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:05:08.220 17:51:01 event.app_repeat -- common/autotest_common.sh@867 -- # local i 00:05:08.220 17:51:01 event.app_repeat -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:05:08.220 17:51:01 event.app_repeat -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:05:08.220 17:51:01 event.app_repeat -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:05:08.220 17:51:01 event.app_repeat -- common/autotest_common.sh@871 -- # break 00:05:08.220 17:51:01 event.app_repeat -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:05:08.220 17:51:01 event.app_repeat -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:05:08.220 17:51:01 event.app_repeat -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:05:08.220 1+0 records in 00:05:08.220 1+0 records out 00:05:08.220 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000176449 s, 23.2 MB/s 00:05:08.220 17:51:01 event.app_repeat -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:05:08.220 17:51:01 event.app_repeat -- common/autotest_common.sh@884 -- # size=4096 00:05:08.220 17:51:01 event.app_repeat -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:05:08.220 17:51:01 event.app_repeat -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:05:08.220 17:51:01 event.app_repeat -- common/autotest_common.sh@887 -- # return 0 00:05:08.220 17:51:01 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:05:08.220 17:51:01 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:05:08.220 17:51:01 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:05:08.481 /dev/nbd1 00:05:08.481 17:51:02 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:05:08.481 17:51:02 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:05:08.481 17:51:02 event.app_repeat -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:05:08.481 17:51:02 event.app_repeat -- common/autotest_common.sh@867 -- # local i 00:05:08.481 17:51:02 event.app_repeat -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:05:08.481 17:51:02 event.app_repeat -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:05:08.481 17:51:02 event.app_repeat -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:05:08.481 17:51:02 event.app_repeat -- common/autotest_common.sh@871 -- # break 00:05:08.481 17:51:02 event.app_repeat -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:05:08.481 17:51:02 event.app_repeat -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:05:08.481 17:51:02 event.app_repeat -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:05:08.481 1+0 records in 00:05:08.481 1+0 records out 00:05:08.482 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000198047 s, 20.7 MB/s 00:05:08.482 17:51:02 event.app_repeat -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:05:08.482 17:51:02 event.app_repeat -- common/autotest_common.sh@884 -- # size=4096 00:05:08.482 17:51:02 event.app_repeat -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:05:08.482 17:51:02 event.app_repeat -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:05:08.482 17:51:02 event.app_repeat -- common/autotest_common.sh@887 -- # return 0 00:05:08.482 17:51:02 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:05:08.482 17:51:02 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:05:08.482 17:51:02 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:05:08.482 17:51:02 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:08.482 17:51:02 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:05:08.740 17:51:02 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:05:08.740 { 00:05:08.740 "nbd_device": "/dev/nbd0", 00:05:08.740 "bdev_name": "Malloc0" 00:05:08.740 }, 00:05:08.740 { 00:05:08.740 "nbd_device": "/dev/nbd1", 00:05:08.740 "bdev_name": "Malloc1" 00:05:08.740 } 00:05:08.740 ]' 00:05:08.740 17:51:02 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:05:08.740 { 00:05:08.740 "nbd_device": "/dev/nbd0", 00:05:08.740 "bdev_name": "Malloc0" 00:05:08.740 }, 00:05:08.740 { 00:05:08.740 "nbd_device": "/dev/nbd1", 00:05:08.740 "bdev_name": "Malloc1" 00:05:08.740 } 00:05:08.740 ]' 00:05:08.740 17:51:02 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:05:08.740 17:51:02 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:05:08.740 /dev/nbd1' 00:05:08.740 17:51:02 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:05:08.740 /dev/nbd1' 00:05:08.740 17:51:02 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:05:08.740 17:51:02 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:05:08.740 17:51:02 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:05:08.740 17:51:02 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:05:08.740 17:51:02 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:05:08.741 256+0 records in 00:05:08.741 256+0 records out 00:05:08.741 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0104584 s, 100 MB/s 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:05:08.741 256+0 records in 00:05:08.741 256+0 records out 00:05:08.741 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0137592 s, 76.2 MB/s 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:05:08.741 256+0 records in 00:05:08.741 256+0 records out 00:05:08.741 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0144121 s, 72.8 MB/s 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:05:08.741 17:51:02 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:05:09.000 17:51:02 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:05:09.000 17:51:02 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:05:09.000 17:51:02 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:05:09.000 17:51:02 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:05:09.000 17:51:02 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:05:09.000 17:51:02 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:05:09.000 17:51:02 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:05:09.000 17:51:02 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:05:09.000 17:51:02 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:05:09.000 17:51:02 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:05:09.259 17:51:02 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:05:09.259 17:51:02 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:05:09.259 17:51:02 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:05:09.259 17:51:02 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:05:09.259 17:51:02 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:05:09.259 17:51:02 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:05:09.259 17:51:02 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:05:09.259 17:51:02 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:05:09.259 17:51:02 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:05:09.259 17:51:02 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:09.259 17:51:02 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:05:09.259 17:51:02 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:05:09.259 17:51:02 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:05:09.259 17:51:02 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:05:09.518 17:51:02 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:05:09.518 17:51:02 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:05:09.518 17:51:02 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:05:09.518 17:51:03 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:05:09.518 17:51:03 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:05:09.518 17:51:03 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:05:09.518 17:51:03 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:05:09.518 17:51:03 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:05:09.518 17:51:03 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:05:09.518 17:51:03 event.app_repeat -- event/event.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:05:09.518 17:51:03 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:05:09.777 [2024-07-15 17:51:03.379514] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:05:09.777 [2024-07-15 17:51:03.447066] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:05:09.777 [2024-07-15 17:51:03.447068] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:09.777 [2024-07-15 17:51:03.488534] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:05:09.777 [2024-07-15 17:51:03.488573] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:05:13.063 17:51:06 event.app_repeat -- event/event.sh@23 -- # for i in {0..2} 00:05:13.063 17:51:06 event.app_repeat -- event/event.sh@24 -- # echo 'spdk_app_start Round 2' 00:05:13.063 spdk_app_start Round 2 00:05:13.063 17:51:06 event.app_repeat -- event/event.sh@25 -- # waitforlisten 412847 /var/tmp/spdk-nbd.sock 00:05:13.063 17:51:06 event.app_repeat -- common/autotest_common.sh@829 -- # '[' -z 412847 ']' 00:05:13.063 17:51:06 event.app_repeat -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:05:13.063 17:51:06 event.app_repeat -- common/autotest_common.sh@834 -- # local max_retries=100 00:05:13.063 17:51:06 event.app_repeat -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:05:13.063 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:05:13.063 17:51:06 event.app_repeat -- common/autotest_common.sh@838 -- # xtrace_disable 00:05:13.063 17:51:06 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:05:13.063 17:51:06 event.app_repeat -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:05:13.063 17:51:06 event.app_repeat -- common/autotest_common.sh@862 -- # return 0 00:05:13.063 17:51:06 event.app_repeat -- event/event.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:05:13.063 Malloc0 00:05:13.063 17:51:06 event.app_repeat -- event/event.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock bdev_malloc_create 64 4096 00:05:13.063 Malloc1 00:05:13.063 17:51:06 event.app_repeat -- event/event.sh@30 -- # nbd_rpc_data_verify /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:05:13.063 17:51:06 event.app_repeat -- bdev/nbd_common.sh@90 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:13.063 17:51:06 event.app_repeat -- bdev/nbd_common.sh@91 -- # bdev_list=('Malloc0' 'Malloc1') 00:05:13.063 17:51:06 event.app_repeat -- bdev/nbd_common.sh@91 -- # local bdev_list 00:05:13.063 17:51:06 event.app_repeat -- bdev/nbd_common.sh@92 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:13.063 17:51:06 event.app_repeat -- bdev/nbd_common.sh@92 -- # local nbd_list 00:05:13.063 17:51:06 event.app_repeat -- bdev/nbd_common.sh@94 -- # nbd_start_disks /var/tmp/spdk-nbd.sock 'Malloc0 Malloc1' '/dev/nbd0 /dev/nbd1' 00:05:13.063 17:51:06 event.app_repeat -- bdev/nbd_common.sh@9 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:13.063 17:51:06 event.app_repeat -- bdev/nbd_common.sh@10 -- # bdev_list=('Malloc0' 'Malloc1') 00:05:13.063 17:51:06 event.app_repeat -- bdev/nbd_common.sh@10 -- # local bdev_list 00:05:13.063 17:51:06 event.app_repeat -- bdev/nbd_common.sh@11 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:13.063 17:51:06 event.app_repeat -- bdev/nbd_common.sh@11 -- # local nbd_list 00:05:13.063 17:51:06 event.app_repeat -- bdev/nbd_common.sh@12 -- # local i 00:05:13.063 17:51:06 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i = 0 )) 00:05:13.063 17:51:06 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:05:13.063 17:51:06 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc0 /dev/nbd0 00:05:13.321 /dev/nbd0 00:05:13.321 17:51:06 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd0 00:05:13.321 17:51:06 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd0 00:05:13.321 17:51:06 event.app_repeat -- common/autotest_common.sh@866 -- # local nbd_name=nbd0 00:05:13.321 17:51:06 event.app_repeat -- common/autotest_common.sh@867 -- # local i 00:05:13.321 17:51:06 event.app_repeat -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:05:13.321 17:51:06 event.app_repeat -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:05:13.321 17:51:06 event.app_repeat -- common/autotest_common.sh@870 -- # grep -q -w nbd0 /proc/partitions 00:05:13.321 17:51:06 event.app_repeat -- common/autotest_common.sh@871 -- # break 00:05:13.321 17:51:06 event.app_repeat -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:05:13.321 17:51:06 event.app_repeat -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:05:13.321 17:51:06 event.app_repeat -- common/autotest_common.sh@883 -- # dd if=/dev/nbd0 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:05:13.321 1+0 records in 00:05:13.321 1+0 records out 00:05:13.321 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000152441 s, 26.9 MB/s 00:05:13.321 17:51:06 event.app_repeat -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:05:13.321 17:51:06 event.app_repeat -- common/autotest_common.sh@884 -- # size=4096 00:05:13.321 17:51:06 event.app_repeat -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:05:13.321 17:51:06 event.app_repeat -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:05:13.321 17:51:06 event.app_repeat -- common/autotest_common.sh@887 -- # return 0 00:05:13.321 17:51:06 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:05:13.321 17:51:06 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:05:13.321 17:51:06 event.app_repeat -- bdev/nbd_common.sh@15 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_start_disk Malloc1 /dev/nbd1 00:05:13.579 /dev/nbd1 00:05:13.579 17:51:07 event.app_repeat -- bdev/nbd_common.sh@17 -- # basename /dev/nbd1 00:05:13.579 17:51:07 event.app_repeat -- bdev/nbd_common.sh@17 -- # waitfornbd nbd1 00:05:13.579 17:51:07 event.app_repeat -- common/autotest_common.sh@866 -- # local nbd_name=nbd1 00:05:13.579 17:51:07 event.app_repeat -- common/autotest_common.sh@867 -- # local i 00:05:13.579 17:51:07 event.app_repeat -- common/autotest_common.sh@869 -- # (( i = 1 )) 00:05:13.579 17:51:07 event.app_repeat -- common/autotest_common.sh@869 -- # (( i <= 20 )) 00:05:13.579 17:51:07 event.app_repeat -- common/autotest_common.sh@870 -- # grep -q -w nbd1 /proc/partitions 00:05:13.579 17:51:07 event.app_repeat -- common/autotest_common.sh@871 -- # break 00:05:13.579 17:51:07 event.app_repeat -- common/autotest_common.sh@882 -- # (( i = 1 )) 00:05:13.579 17:51:07 event.app_repeat -- common/autotest_common.sh@882 -- # (( i <= 20 )) 00:05:13.579 17:51:07 event.app_repeat -- common/autotest_common.sh@883 -- # dd if=/dev/nbd1 of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest bs=4096 count=1 iflag=direct 00:05:13.579 1+0 records in 00:05:13.579 1+0 records out 00:05:13.579 4096 bytes (4.1 kB, 4.0 KiB) copied, 0.000181481 s, 22.6 MB/s 00:05:13.579 17:51:07 event.app_repeat -- common/autotest_common.sh@884 -- # stat -c %s /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:05:13.579 17:51:07 event.app_repeat -- common/autotest_common.sh@884 -- # size=4096 00:05:13.579 17:51:07 event.app_repeat -- common/autotest_common.sh@885 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdtest 00:05:13.579 17:51:07 event.app_repeat -- common/autotest_common.sh@886 -- # '[' 4096 '!=' 0 ']' 00:05:13.579 17:51:07 event.app_repeat -- common/autotest_common.sh@887 -- # return 0 00:05:13.579 17:51:07 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i++ )) 00:05:13.579 17:51:07 event.app_repeat -- bdev/nbd_common.sh@14 -- # (( i < 2 )) 00:05:13.579 17:51:07 event.app_repeat -- bdev/nbd_common.sh@95 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:05:13.579 17:51:07 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:13.579 17:51:07 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[ 00:05:13.837 { 00:05:13.837 "nbd_device": "/dev/nbd0", 00:05:13.837 "bdev_name": "Malloc0" 00:05:13.837 }, 00:05:13.837 { 00:05:13.837 "nbd_device": "/dev/nbd1", 00:05:13.837 "bdev_name": "Malloc1" 00:05:13.837 } 00:05:13.837 ]' 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[ 00:05:13.837 { 00:05:13.837 "nbd_device": "/dev/nbd0", 00:05:13.837 "bdev_name": "Malloc0" 00:05:13.837 }, 00:05:13.837 { 00:05:13.837 "nbd_device": "/dev/nbd1", 00:05:13.837 "bdev_name": "Malloc1" 00:05:13.837 } 00:05:13.837 ]' 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name='/dev/nbd0 00:05:13.837 /dev/nbd1' 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '/dev/nbd0 00:05:13.837 /dev/nbd1' 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=2 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 2 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@95 -- # count=2 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@96 -- # '[' 2 -ne 2 ']' 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@100 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' write 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=write 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' write = write ']' 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@76 -- # dd if=/dev/urandom of=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest bs=4096 count=256 00:05:13.837 256+0 records in 00:05:13.837 256+0 records out 00:05:13.837 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.00397313 s, 264 MB/s 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd0 bs=4096 count=256 oflag=direct 00:05:13.837 256+0 records in 00:05:13.837 256+0 records out 00:05:13.837 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.0138791 s, 75.6 MB/s 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@77 -- # for i in "${nbd_list[@]}" 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@78 -- # dd if=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest of=/dev/nbd1 bs=4096 count=256 oflag=direct 00:05:13.837 256+0 records in 00:05:13.837 256+0 records out 00:05:13.837 1048576 bytes (1.0 MB, 1.0 MiB) copied, 0.014489 s, 72.4 MB/s 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@101 -- # nbd_dd_data_verify '/dev/nbd0 /dev/nbd1' verify 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@70 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@70 -- # local nbd_list 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@71 -- # local operation=verify 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@72 -- # local tmp_file=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@74 -- # '[' verify = write ']' 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@80 -- # '[' verify = verify ']' 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd0 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@82 -- # for i in "${nbd_list[@]}" 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@83 -- # cmp -b -n 1M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest /dev/nbd1 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@85 -- # rm /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/nbdrandtest 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@103 -- # nbd_stop_disks /var/tmp/spdk-nbd.sock '/dev/nbd0 /dev/nbd1' 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@49 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@50 -- # nbd_list=('/dev/nbd0' '/dev/nbd1') 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@50 -- # local nbd_list 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@51 -- # local i 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:05:13.837 17:51:07 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd0 00:05:14.095 17:51:07 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd0 00:05:14.095 17:51:07 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd0 00:05:14.095 17:51:07 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd0 00:05:14.095 17:51:07 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:05:14.095 17:51:07 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:05:14.095 17:51:07 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd0 /proc/partitions 00:05:14.095 17:51:07 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:05:14.095 17:51:07 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:05:14.095 17:51:07 event.app_repeat -- bdev/nbd_common.sh@53 -- # for i in "${nbd_list[@]}" 00:05:14.095 17:51:07 event.app_repeat -- bdev/nbd_common.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_stop_disk /dev/nbd1 00:05:14.353 17:51:07 event.app_repeat -- bdev/nbd_common.sh@55 -- # basename /dev/nbd1 00:05:14.353 17:51:07 event.app_repeat -- bdev/nbd_common.sh@55 -- # waitfornbd_exit nbd1 00:05:14.353 17:51:07 event.app_repeat -- bdev/nbd_common.sh@35 -- # local nbd_name=nbd1 00:05:14.353 17:51:07 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i = 1 )) 00:05:14.353 17:51:07 event.app_repeat -- bdev/nbd_common.sh@37 -- # (( i <= 20 )) 00:05:14.353 17:51:07 event.app_repeat -- bdev/nbd_common.sh@38 -- # grep -q -w nbd1 /proc/partitions 00:05:14.353 17:51:07 event.app_repeat -- bdev/nbd_common.sh@41 -- # break 00:05:14.353 17:51:07 event.app_repeat -- bdev/nbd_common.sh@45 -- # return 0 00:05:14.353 17:51:07 event.app_repeat -- bdev/nbd_common.sh@104 -- # nbd_get_count /var/tmp/spdk-nbd.sock 00:05:14.353 17:51:07 event.app_repeat -- bdev/nbd_common.sh@61 -- # local rpc_server=/var/tmp/spdk-nbd.sock 00:05:14.353 17:51:07 event.app_repeat -- bdev/nbd_common.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock nbd_get_disks 00:05:14.353 17:51:08 event.app_repeat -- bdev/nbd_common.sh@63 -- # nbd_disks_json='[]' 00:05:14.353 17:51:08 event.app_repeat -- bdev/nbd_common.sh@64 -- # echo '[]' 00:05:14.353 17:51:08 event.app_repeat -- bdev/nbd_common.sh@64 -- # jq -r '.[] | .nbd_device' 00:05:14.353 17:51:08 event.app_repeat -- bdev/nbd_common.sh@64 -- # nbd_disks_name= 00:05:14.353 17:51:08 event.app_repeat -- bdev/nbd_common.sh@65 -- # echo '' 00:05:14.353 17:51:08 event.app_repeat -- bdev/nbd_common.sh@65 -- # grep -c /dev/nbd 00:05:14.353 17:51:08 event.app_repeat -- bdev/nbd_common.sh@65 -- # true 00:05:14.353 17:51:08 event.app_repeat -- bdev/nbd_common.sh@65 -- # count=0 00:05:14.353 17:51:08 event.app_repeat -- bdev/nbd_common.sh@66 -- # echo 0 00:05:14.353 17:51:08 event.app_repeat -- bdev/nbd_common.sh@104 -- # count=0 00:05:14.353 17:51:08 event.app_repeat -- bdev/nbd_common.sh@105 -- # '[' 0 -ne 0 ']' 00:05:14.353 17:51:08 event.app_repeat -- bdev/nbd_common.sh@109 -- # return 0 00:05:14.353 17:51:08 event.app_repeat -- event/event.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/spdk-nbd.sock spdk_kill_instance SIGTERM 00:05:14.610 17:51:08 event.app_repeat -- event/event.sh@35 -- # sleep 3 00:05:14.869 [2024-07-15 17:51:08.461033] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:05:14.869 [2024-07-15 17:51:08.533281] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:05:14.869 [2024-07-15 17:51:08.533284] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:14.869 [2024-07-15 17:51:08.574088] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_register' already registered. 00:05:14.869 [2024-07-15 17:51:08.574128] notify.c: 45:spdk_notify_type_register: *NOTICE*: Notification type 'bdev_unregister' already registered. 00:05:18.161 17:51:11 event.app_repeat -- event/event.sh@38 -- # waitforlisten 412847 /var/tmp/spdk-nbd.sock 00:05:18.161 17:51:11 event.app_repeat -- common/autotest_common.sh@829 -- # '[' -z 412847 ']' 00:05:18.161 17:51:11 event.app_repeat -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk-nbd.sock 00:05:18.161 17:51:11 event.app_repeat -- common/autotest_common.sh@834 -- # local max_retries=100 00:05:18.161 17:51:11 event.app_repeat -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock...' 00:05:18.161 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk-nbd.sock... 00:05:18.161 17:51:11 event.app_repeat -- common/autotest_common.sh@838 -- # xtrace_disable 00:05:18.161 17:51:11 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:05:18.161 17:51:11 event.app_repeat -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:05:18.161 17:51:11 event.app_repeat -- common/autotest_common.sh@862 -- # return 0 00:05:18.161 17:51:11 event.app_repeat -- event/event.sh@39 -- # killprocess 412847 00:05:18.161 17:51:11 event.app_repeat -- common/autotest_common.sh@948 -- # '[' -z 412847 ']' 00:05:18.161 17:51:11 event.app_repeat -- common/autotest_common.sh@952 -- # kill -0 412847 00:05:18.161 17:51:11 event.app_repeat -- common/autotest_common.sh@953 -- # uname 00:05:18.161 17:51:11 event.app_repeat -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:05:18.161 17:51:11 event.app_repeat -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 412847 00:05:18.161 17:51:11 event.app_repeat -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:05:18.161 17:51:11 event.app_repeat -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:05:18.161 17:51:11 event.app_repeat -- common/autotest_common.sh@966 -- # echo 'killing process with pid 412847' 00:05:18.161 killing process with pid 412847 00:05:18.161 17:51:11 event.app_repeat -- common/autotest_common.sh@967 -- # kill 412847 00:05:18.161 17:51:11 event.app_repeat -- common/autotest_common.sh@972 -- # wait 412847 00:05:18.161 spdk_app_start is called in Round 0. 00:05:18.161 Shutdown signal received, stop current app iteration 00:05:18.161 Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 reinitialization... 00:05:18.161 spdk_app_start is called in Round 1. 00:05:18.161 Shutdown signal received, stop current app iteration 00:05:18.161 Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 reinitialization... 00:05:18.161 spdk_app_start is called in Round 2. 00:05:18.161 Shutdown signal received, stop current app iteration 00:05:18.161 Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 reinitialization... 00:05:18.161 spdk_app_start is called in Round 3. 00:05:18.161 Shutdown signal received, stop current app iteration 00:05:18.161 17:51:11 event.app_repeat -- event/event.sh@40 -- # trap - SIGINT SIGTERM EXIT 00:05:18.161 17:51:11 event.app_repeat -- event/event.sh@42 -- # return 0 00:05:18.161 00:05:18.161 real 0m15.678s 00:05:18.161 user 0m33.929s 00:05:18.161 sys 0m2.340s 00:05:18.161 17:51:11 event.app_repeat -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:18.161 17:51:11 event.app_repeat -- common/autotest_common.sh@10 -- # set +x 00:05:18.161 ************************************ 00:05:18.161 END TEST app_repeat 00:05:18.161 ************************************ 00:05:18.161 17:51:11 event -- common/autotest_common.sh@1142 -- # return 0 00:05:18.161 17:51:11 event -- event/event.sh@54 -- # (( SPDK_TEST_CRYPTO == 0 )) 00:05:18.161 17:51:11 event -- event/event.sh@55 -- # run_test cpu_locks /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/cpu_locks.sh 00:05:18.161 17:51:11 event -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:05:18.161 17:51:11 event -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:18.161 17:51:11 event -- common/autotest_common.sh@10 -- # set +x 00:05:18.161 ************************************ 00:05:18.161 START TEST cpu_locks 00:05:18.161 ************************************ 00:05:18.161 17:51:11 event.cpu_locks -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event/cpu_locks.sh 00:05:18.161 * Looking for test storage... 00:05:18.161 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/event 00:05:18.161 17:51:11 event.cpu_locks -- event/cpu_locks.sh@11 -- # rpc_sock1=/var/tmp/spdk.sock 00:05:18.161 17:51:11 event.cpu_locks -- event/cpu_locks.sh@12 -- # rpc_sock2=/var/tmp/spdk2.sock 00:05:18.161 17:51:11 event.cpu_locks -- event/cpu_locks.sh@164 -- # trap cleanup EXIT SIGTERM SIGINT 00:05:18.161 17:51:11 event.cpu_locks -- event/cpu_locks.sh@166 -- # run_test default_locks default_locks 00:05:18.161 17:51:11 event.cpu_locks -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:05:18.161 17:51:11 event.cpu_locks -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:18.161 17:51:11 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:05:18.161 ************************************ 00:05:18.161 START TEST default_locks 00:05:18.161 ************************************ 00:05:18.161 17:51:11 event.cpu_locks.default_locks -- common/autotest_common.sh@1123 -- # default_locks 00:05:18.161 17:51:11 event.cpu_locks.default_locks -- event/cpu_locks.sh@46 -- # spdk_tgt_pid=416304 00:05:18.161 17:51:11 event.cpu_locks.default_locks -- event/cpu_locks.sh@47 -- # waitforlisten 416304 00:05:18.161 17:51:11 event.cpu_locks.default_locks -- common/autotest_common.sh@829 -- # '[' -z 416304 ']' 00:05:18.161 17:51:11 event.cpu_locks.default_locks -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:18.161 17:51:11 event.cpu_locks.default_locks -- common/autotest_common.sh@834 -- # local max_retries=100 00:05:18.161 17:51:11 event.cpu_locks.default_locks -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:18.161 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:18.161 17:51:11 event.cpu_locks.default_locks -- event/cpu_locks.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:05:18.161 17:51:11 event.cpu_locks.default_locks -- common/autotest_common.sh@838 -- # xtrace_disable 00:05:18.161 17:51:11 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:05:18.421 [2024-07-15 17:51:11.899185] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:18.421 [2024-07-15 17:51:11.899234] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid416304 ] 00:05:18.421 EAL: No free 2048 kB hugepages reported on node 1 00:05:18.421 [2024-07-15 17:51:11.951777] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:18.421 [2024-07-15 17:51:12.031436] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:18.990 17:51:12 event.cpu_locks.default_locks -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:05:18.990 17:51:12 event.cpu_locks.default_locks -- common/autotest_common.sh@862 -- # return 0 00:05:18.990 17:51:12 event.cpu_locks.default_locks -- event/cpu_locks.sh@49 -- # locks_exist 416304 00:05:18.990 17:51:12 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # lslocks -p 416304 00:05:18.990 17:51:12 event.cpu_locks.default_locks -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:05:19.560 lslocks: write error 00:05:19.560 17:51:12 event.cpu_locks.default_locks -- event/cpu_locks.sh@50 -- # killprocess 416304 00:05:19.560 17:51:12 event.cpu_locks.default_locks -- common/autotest_common.sh@948 -- # '[' -z 416304 ']' 00:05:19.560 17:51:12 event.cpu_locks.default_locks -- common/autotest_common.sh@952 -- # kill -0 416304 00:05:19.560 17:51:12 event.cpu_locks.default_locks -- common/autotest_common.sh@953 -- # uname 00:05:19.560 17:51:12 event.cpu_locks.default_locks -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:05:19.560 17:51:12 event.cpu_locks.default_locks -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 416304 00:05:19.560 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:05:19.560 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:05:19.560 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@966 -- # echo 'killing process with pid 416304' 00:05:19.560 killing process with pid 416304 00:05:19.560 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@967 -- # kill 416304 00:05:19.560 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@972 -- # wait 416304 00:05:19.819 17:51:13 event.cpu_locks.default_locks -- event/cpu_locks.sh@52 -- # NOT waitforlisten 416304 00:05:19.819 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@648 -- # local es=0 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@650 -- # valid_exec_arg waitforlisten 416304 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@636 -- # local arg=waitforlisten 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@640 -- # type -t waitforlisten 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@651 -- # waitforlisten 416304 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@829 -- # '[' -z 416304 ']' 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@834 -- # local max_retries=100 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:19.820 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@838 -- # xtrace_disable 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:05:19.820 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 844: kill: (416304) - No such process 00:05:19.820 ERROR: process (pid: 416304) is no longer running 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@862 -- # return 1 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@651 -- # es=1 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- event/cpu_locks.sh@54 -- # no_locks 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # lock_files=() 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- event/cpu_locks.sh@26 -- # local lock_files 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:05:19.820 00:05:19.820 real 0m1.502s 00:05:19.820 user 0m1.583s 00:05:19.820 sys 0m0.473s 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:19.820 17:51:13 event.cpu_locks.default_locks -- common/autotest_common.sh@10 -- # set +x 00:05:19.820 ************************************ 00:05:19.820 END TEST default_locks 00:05:19.820 ************************************ 00:05:19.820 17:51:13 event.cpu_locks -- common/autotest_common.sh@1142 -- # return 0 00:05:19.820 17:51:13 event.cpu_locks -- event/cpu_locks.sh@167 -- # run_test default_locks_via_rpc default_locks_via_rpc 00:05:19.820 17:51:13 event.cpu_locks -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:05:19.820 17:51:13 event.cpu_locks -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:19.820 17:51:13 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:05:19.820 ************************************ 00:05:19.820 START TEST default_locks_via_rpc 00:05:19.820 ************************************ 00:05:19.820 17:51:13 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1123 -- # default_locks_via_rpc 00:05:19.820 17:51:13 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:05:19.820 17:51:13 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@62 -- # spdk_tgt_pid=416567 00:05:19.820 17:51:13 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@63 -- # waitforlisten 416567 00:05:19.820 17:51:13 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@829 -- # '[' -z 416567 ']' 00:05:19.820 17:51:13 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:19.820 17:51:13 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@834 -- # local max_retries=100 00:05:19.820 17:51:13 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:19.820 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:19.820 17:51:13 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@838 -- # xtrace_disable 00:05:19.820 17:51:13 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:19.820 [2024-07-15 17:51:13.458059] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:19.820 [2024-07-15 17:51:13.458098] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid416567 ] 00:05:19.820 EAL: No free 2048 kB hugepages reported on node 1 00:05:19.820 [2024-07-15 17:51:13.511174] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:20.078 [2024-07-15 17:51:13.590694] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:20.646 17:51:14 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:05:20.646 17:51:14 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@862 -- # return 0 00:05:20.646 17:51:14 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@65 -- # rpc_cmd framework_disable_cpumask_locks 00:05:20.646 17:51:14 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:05:20.646 17:51:14 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:20.646 17:51:14 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:05:20.646 17:51:14 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@67 -- # no_locks 00:05:20.646 17:51:14 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # lock_files=() 00:05:20.646 17:51:14 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@26 -- # local lock_files 00:05:20.646 17:51:14 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@27 -- # (( 0 != 0 )) 00:05:20.646 17:51:14 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@69 -- # rpc_cmd framework_enable_cpumask_locks 00:05:20.646 17:51:14 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:05:20.646 17:51:14 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:20.646 17:51:14 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:05:20.646 17:51:14 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@71 -- # locks_exist 416567 00:05:20.646 17:51:14 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # lslocks -p 416567 00:05:20.646 17:51:14 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:05:20.905 17:51:14 event.cpu_locks.default_locks_via_rpc -- event/cpu_locks.sh@73 -- # killprocess 416567 00:05:20.905 17:51:14 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@948 -- # '[' -z 416567 ']' 00:05:20.905 17:51:14 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@952 -- # kill -0 416567 00:05:20.905 17:51:14 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@953 -- # uname 00:05:20.905 17:51:14 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:05:20.905 17:51:14 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 416567 00:05:21.164 17:51:14 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:05:21.164 17:51:14 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:05:21.164 17:51:14 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@966 -- # echo 'killing process with pid 416567' 00:05:21.164 killing process with pid 416567 00:05:21.164 17:51:14 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@967 -- # kill 416567 00:05:21.164 17:51:14 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@972 -- # wait 416567 00:05:21.424 00:05:21.424 real 0m1.524s 00:05:21.424 user 0m1.620s 00:05:21.424 sys 0m0.483s 00:05:21.424 17:51:14 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:21.424 17:51:14 event.cpu_locks.default_locks_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:21.424 ************************************ 00:05:21.424 END TEST default_locks_via_rpc 00:05:21.424 ************************************ 00:05:21.424 17:51:14 event.cpu_locks -- common/autotest_common.sh@1142 -- # return 0 00:05:21.424 17:51:14 event.cpu_locks -- event/cpu_locks.sh@168 -- # run_test non_locking_app_on_locked_coremask non_locking_app_on_locked_coremask 00:05:21.424 17:51:14 event.cpu_locks -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:05:21.424 17:51:14 event.cpu_locks -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:21.424 17:51:14 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:05:21.424 ************************************ 00:05:21.424 START TEST non_locking_app_on_locked_coremask 00:05:21.424 ************************************ 00:05:21.424 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1123 -- # non_locking_app_on_locked_coremask 00:05:21.424 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@80 -- # spdk_tgt_pid=416833 00:05:21.424 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@81 -- # waitforlisten 416833 /var/tmp/spdk.sock 00:05:21.424 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@829 -- # '[' -z 416833 ']' 00:05:21.424 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:21.424 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@834 -- # local max_retries=100 00:05:21.424 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:21.424 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:21.424 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:05:21.424 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # xtrace_disable 00:05:21.424 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:21.424 [2024-07-15 17:51:15.048634] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:21.424 [2024-07-15 17:51:15.048673] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid416833 ] 00:05:21.424 EAL: No free 2048 kB hugepages reported on node 1 00:05:21.424 [2024-07-15 17:51:15.102733] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:21.684 [2024-07-15 17:51:15.182577] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:22.253 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:05:22.253 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@862 -- # return 0 00:05:22.253 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@84 -- # spdk_tgt_pid2=417061 00:05:22.253 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@85 -- # waitforlisten 417061 /var/tmp/spdk2.sock 00:05:22.253 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@829 -- # '[' -z 417061 ']' 00:05:22.253 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk2.sock 00:05:22.253 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@834 -- # local max_retries=100 00:05:22.253 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:05:22.253 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:05:22.253 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # xtrace_disable 00:05:22.253 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks -r /var/tmp/spdk2.sock 00:05:22.253 17:51:15 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:22.253 [2024-07-15 17:51:15.892463] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:22.253 [2024-07-15 17:51:15.892509] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid417061 ] 00:05:22.253 EAL: No free 2048 kB hugepages reported on node 1 00:05:22.253 [2024-07-15 17:51:15.967173] app.c: 906:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:05:22.253 [2024-07-15 17:51:15.967198] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:22.512 [2024-07-15 17:51:16.113543] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:23.081 17:51:16 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:05:23.081 17:51:16 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@862 -- # return 0 00:05:23.081 17:51:16 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@87 -- # locks_exist 416833 00:05:23.081 17:51:16 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 416833 00:05:23.081 17:51:16 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:05:23.648 lslocks: write error 00:05:23.648 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@89 -- # killprocess 416833 00:05:23.648 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@948 -- # '[' -z 416833 ']' 00:05:23.648 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@952 -- # kill -0 416833 00:05:23.648 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@953 -- # uname 00:05:23.648 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:05:23.648 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 416833 00:05:23.648 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:05:23.648 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:05:23.648 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@966 -- # echo 'killing process with pid 416833' 00:05:23.648 killing process with pid 416833 00:05:23.648 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@967 -- # kill 416833 00:05:23.648 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # wait 416833 00:05:24.215 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- event/cpu_locks.sh@90 -- # killprocess 417061 00:05:24.216 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@948 -- # '[' -z 417061 ']' 00:05:24.216 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@952 -- # kill -0 417061 00:05:24.216 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@953 -- # uname 00:05:24.475 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:05:24.475 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 417061 00:05:24.475 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:05:24.475 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:05:24.475 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@966 -- # echo 'killing process with pid 417061' 00:05:24.475 killing process with pid 417061 00:05:24.475 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@967 -- # kill 417061 00:05:24.475 17:51:17 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # wait 417061 00:05:24.735 00:05:24.735 real 0m3.294s 00:05:24.735 user 0m3.512s 00:05:24.735 sys 0m0.934s 00:05:24.735 17:51:18 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:24.735 17:51:18 event.cpu_locks.non_locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:24.735 ************************************ 00:05:24.735 END TEST non_locking_app_on_locked_coremask 00:05:24.735 ************************************ 00:05:24.735 17:51:18 event.cpu_locks -- common/autotest_common.sh@1142 -- # return 0 00:05:24.735 17:51:18 event.cpu_locks -- event/cpu_locks.sh@169 -- # run_test locking_app_on_unlocked_coremask locking_app_on_unlocked_coremask 00:05:24.735 17:51:18 event.cpu_locks -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:05:24.735 17:51:18 event.cpu_locks -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:24.735 17:51:18 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:05:24.735 ************************************ 00:05:24.735 START TEST locking_app_on_unlocked_coremask 00:05:24.735 ************************************ 00:05:24.735 17:51:18 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1123 -- # locking_app_on_unlocked_coremask 00:05:24.735 17:51:18 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@98 -- # spdk_tgt_pid=417549 00:05:24.735 17:51:18 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@99 -- # waitforlisten 417549 /var/tmp/spdk.sock 00:05:24.735 17:51:18 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@97 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 --disable-cpumask-locks 00:05:24.735 17:51:18 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@829 -- # '[' -z 417549 ']' 00:05:24.735 17:51:18 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:24.735 17:51:18 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@834 -- # local max_retries=100 00:05:24.735 17:51:18 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:24.735 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:24.735 17:51:18 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@838 -- # xtrace_disable 00:05:24.735 17:51:18 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:24.735 [2024-07-15 17:51:18.395541] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:24.735 [2024-07-15 17:51:18.395576] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid417549 ] 00:05:24.735 EAL: No free 2048 kB hugepages reported on node 1 00:05:24.735 [2024-07-15 17:51:18.449100] app.c: 906:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:05:24.735 [2024-07-15 17:51:18.449123] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:24.994 [2024-07-15 17:51:18.528632] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:25.564 17:51:19 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:05:25.564 17:51:19 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@862 -- # return 0 00:05:25.564 17:51:19 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@102 -- # spdk_tgt_pid2=417589 00:05:25.564 17:51:19 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@101 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:05:25.564 17:51:19 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@103 -- # waitforlisten 417589 /var/tmp/spdk2.sock 00:05:25.564 17:51:19 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@829 -- # '[' -z 417589 ']' 00:05:25.564 17:51:19 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk2.sock 00:05:25.564 17:51:19 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@834 -- # local max_retries=100 00:05:25.564 17:51:19 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:05:25.564 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:05:25.564 17:51:19 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@838 -- # xtrace_disable 00:05:25.564 17:51:19 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:25.564 [2024-07-15 17:51:19.226749] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:25.564 [2024-07-15 17:51:19.226807] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid417589 ] 00:05:25.564 EAL: No free 2048 kB hugepages reported on node 1 00:05:25.860 [2024-07-15 17:51:19.301136] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:25.860 [2024-07-15 17:51:19.457655] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:26.429 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:05:26.430 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@862 -- # return 0 00:05:26.430 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@105 -- # locks_exist 417589 00:05:26.430 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 417589 00:05:26.430 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:05:26.689 lslocks: write error 00:05:26.689 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@107 -- # killprocess 417549 00:05:26.689 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@948 -- # '[' -z 417549 ']' 00:05:26.689 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@952 -- # kill -0 417549 00:05:26.689 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@953 -- # uname 00:05:26.689 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:05:26.689 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 417549 00:05:26.689 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:05:26.689 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:05:26.689 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@966 -- # echo 'killing process with pid 417549' 00:05:26.689 killing process with pid 417549 00:05:26.689 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@967 -- # kill 417549 00:05:26.689 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@972 -- # wait 417549 00:05:27.258 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- event/cpu_locks.sh@108 -- # killprocess 417589 00:05:27.258 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@948 -- # '[' -z 417589 ']' 00:05:27.258 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@952 -- # kill -0 417589 00:05:27.258 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@953 -- # uname 00:05:27.258 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:05:27.258 17:51:20 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 417589 00:05:27.517 17:51:21 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:05:27.517 17:51:21 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:05:27.517 17:51:21 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@966 -- # echo 'killing process with pid 417589' 00:05:27.517 killing process with pid 417589 00:05:27.517 17:51:21 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@967 -- # kill 417589 00:05:27.517 17:51:21 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@972 -- # wait 417589 00:05:27.776 00:05:27.776 real 0m2.959s 00:05:27.776 user 0m3.172s 00:05:27.776 sys 0m0.795s 00:05:27.777 17:51:21 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:27.777 17:51:21 event.cpu_locks.locking_app_on_unlocked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:27.777 ************************************ 00:05:27.777 END TEST locking_app_on_unlocked_coremask 00:05:27.777 ************************************ 00:05:27.777 17:51:21 event.cpu_locks -- common/autotest_common.sh@1142 -- # return 0 00:05:27.777 17:51:21 event.cpu_locks -- event/cpu_locks.sh@170 -- # run_test locking_app_on_locked_coremask locking_app_on_locked_coremask 00:05:27.777 17:51:21 event.cpu_locks -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:05:27.777 17:51:21 event.cpu_locks -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:27.777 17:51:21 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:05:27.777 ************************************ 00:05:27.777 START TEST locking_app_on_locked_coremask 00:05:27.777 ************************************ 00:05:27.777 17:51:21 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1123 -- # locking_app_on_locked_coremask 00:05:27.777 17:51:21 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@115 -- # spdk_tgt_pid=418058 00:05:27.777 17:51:21 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@116 -- # waitforlisten 418058 /var/tmp/spdk.sock 00:05:27.777 17:51:21 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@829 -- # '[' -z 418058 ']' 00:05:27.777 17:51:21 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:27.777 17:51:21 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@834 -- # local max_retries=100 00:05:27.777 17:51:21 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:27.777 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:27.777 17:51:21 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # xtrace_disable 00:05:27.777 17:51:21 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:27.777 17:51:21 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@114 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 00:05:27.777 [2024-07-15 17:51:21.422637] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:27.777 [2024-07-15 17:51:21.422676] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid418058 ] 00:05:27.777 EAL: No free 2048 kB hugepages reported on node 1 00:05:27.777 [2024-07-15 17:51:21.475213] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:28.034 [2024-07-15 17:51:21.555082] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:28.603 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:05:28.603 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@862 -- # return 0 00:05:28.603 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@119 -- # spdk_tgt_pid2=418263 00:05:28.603 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@120 -- # NOT waitforlisten 418263 /var/tmp/spdk2.sock 00:05:28.603 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@648 -- # local es=0 00:05:28.603 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@650 -- # valid_exec_arg waitforlisten 418263 /var/tmp/spdk2.sock 00:05:28.603 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@636 -- # local arg=waitforlisten 00:05:28.603 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@118 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1 -r /var/tmp/spdk2.sock 00:05:28.603 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:05:28.603 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@640 -- # type -t waitforlisten 00:05:28.603 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:05:28.603 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@651 -- # waitforlisten 418263 /var/tmp/spdk2.sock 00:05:28.603 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@829 -- # '[' -z 418263 ']' 00:05:28.603 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk2.sock 00:05:28.603 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@834 -- # local max_retries=100 00:05:28.603 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:05:28.603 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:05:28.603 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@838 -- # xtrace_disable 00:05:28.603 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:28.603 [2024-07-15 17:51:22.254111] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:28.603 [2024-07-15 17:51:22.254159] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid418263 ] 00:05:28.603 EAL: No free 2048 kB hugepages reported on node 1 00:05:28.603 [2024-07-15 17:51:22.330577] app.c: 771:claim_cpu_cores: *ERROR*: Cannot create lock on core 0, probably process 418058 has claimed it. 00:05:28.603 [2024-07-15 17:51:22.330612] app.c: 902:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:05:29.171 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 844: kill: (418263) - No such process 00:05:29.171 ERROR: process (pid: 418263) is no longer running 00:05:29.171 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:05:29.171 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@862 -- # return 1 00:05:29.171 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@651 -- # es=1 00:05:29.171 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:05:29.171 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:05:29.171 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:05:29.171 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@122 -- # locks_exist 418058 00:05:29.171 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # lslocks -p 418058 00:05:29.171 17:51:22 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@22 -- # grep -q spdk_cpu_lock 00:05:29.431 lslocks: write error 00:05:29.431 17:51:23 event.cpu_locks.locking_app_on_locked_coremask -- event/cpu_locks.sh@124 -- # killprocess 418058 00:05:29.431 17:51:23 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@948 -- # '[' -z 418058 ']' 00:05:29.431 17:51:23 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@952 -- # kill -0 418058 00:05:29.431 17:51:23 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@953 -- # uname 00:05:29.431 17:51:23 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:05:29.431 17:51:23 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 418058 00:05:29.431 17:51:23 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:05:29.431 17:51:23 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:05:29.431 17:51:23 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@966 -- # echo 'killing process with pid 418058' 00:05:29.431 killing process with pid 418058 00:05:29.431 17:51:23 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@967 -- # kill 418058 00:05:29.431 17:51:23 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@972 -- # wait 418058 00:05:29.998 00:05:29.998 real 0m2.067s 00:05:29.998 user 0m2.276s 00:05:29.998 sys 0m0.514s 00:05:29.998 17:51:23 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:29.998 17:51:23 event.cpu_locks.locking_app_on_locked_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:29.998 ************************************ 00:05:29.998 END TEST locking_app_on_locked_coremask 00:05:29.998 ************************************ 00:05:29.998 17:51:23 event.cpu_locks -- common/autotest_common.sh@1142 -- # return 0 00:05:29.998 17:51:23 event.cpu_locks -- event/cpu_locks.sh@171 -- # run_test locking_overlapped_coremask locking_overlapped_coremask 00:05:29.998 17:51:23 event.cpu_locks -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:05:29.998 17:51:23 event.cpu_locks -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:29.998 17:51:23 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:05:29.998 ************************************ 00:05:29.998 START TEST locking_overlapped_coremask 00:05:29.998 ************************************ 00:05:29.998 17:51:23 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1123 -- # locking_overlapped_coremask 00:05:29.998 17:51:23 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@132 -- # spdk_tgt_pid=418470 00:05:29.998 17:51:23 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@133 -- # waitforlisten 418470 /var/tmp/spdk.sock 00:05:29.998 17:51:23 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@829 -- # '[' -z 418470 ']' 00:05:29.998 17:51:23 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:29.998 17:51:23 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@834 -- # local max_retries=100 00:05:29.998 17:51:23 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@131 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x7 00:05:29.998 17:51:23 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:29.998 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:29.998 17:51:23 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@838 -- # xtrace_disable 00:05:29.998 17:51:23 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:29.998 [2024-07-15 17:51:23.543997] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:29.998 [2024-07-15 17:51:23.544035] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid418470 ] 00:05:29.998 EAL: No free 2048 kB hugepages reported on node 1 00:05:29.998 [2024-07-15 17:51:23.597879] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:05:29.998 [2024-07-15 17:51:23.678705] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:05:29.998 [2024-07-15 17:51:23.678723] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:05:29.998 [2024-07-15 17:51:23.678725] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:30.936 17:51:24 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:05:30.936 17:51:24 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@862 -- # return 0 00:05:30.936 17:51:24 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@136 -- # spdk_tgt_pid2=418562 00:05:30.936 17:51:24 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@137 -- # NOT waitforlisten 418562 /var/tmp/spdk2.sock 00:05:30.936 17:51:24 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@135 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock 00:05:30.937 17:51:24 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@648 -- # local es=0 00:05:30.937 17:51:24 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@650 -- # valid_exec_arg waitforlisten 418562 /var/tmp/spdk2.sock 00:05:30.937 17:51:24 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@636 -- # local arg=waitforlisten 00:05:30.937 17:51:24 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:05:30.937 17:51:24 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@640 -- # type -t waitforlisten 00:05:30.937 17:51:24 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:05:30.937 17:51:24 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@651 -- # waitforlisten 418562 /var/tmp/spdk2.sock 00:05:30.937 17:51:24 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@829 -- # '[' -z 418562 ']' 00:05:30.937 17:51:24 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk2.sock 00:05:30.937 17:51:24 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@834 -- # local max_retries=100 00:05:30.937 17:51:24 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:05:30.937 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:05:30.937 17:51:24 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@838 -- # xtrace_disable 00:05:30.937 17:51:24 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:30.937 [2024-07-15 17:51:24.403664] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:30.937 [2024-07-15 17:51:24.403713] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid418562 ] 00:05:30.937 EAL: No free 2048 kB hugepages reported on node 1 00:05:30.937 [2024-07-15 17:51:24.479260] app.c: 771:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 418470 has claimed it. 00:05:30.937 [2024-07-15 17:51:24.479296] app.c: 902:spdk_app_start: *ERROR*: Unable to acquire lock on assigned core mask - exiting. 00:05:31.506 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 844: kill: (418562) - No such process 00:05:31.506 ERROR: process (pid: 418562) is no longer running 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@862 -- # return 1 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@651 -- # es=1 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@139 -- # check_remaining_locks 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- event/cpu_locks.sh@141 -- # killprocess 418470 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@948 -- # '[' -z 418470 ']' 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@952 -- # kill -0 418470 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@953 -- # uname 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 418470 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@966 -- # echo 'killing process with pid 418470' 00:05:31.506 killing process with pid 418470 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@967 -- # kill 418470 00:05:31.506 17:51:25 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@972 -- # wait 418470 00:05:31.766 00:05:31.766 real 0m1.883s 00:05:31.766 user 0m5.336s 00:05:31.766 sys 0m0.394s 00:05:31.766 17:51:25 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:31.766 17:51:25 event.cpu_locks.locking_overlapped_coremask -- common/autotest_common.sh@10 -- # set +x 00:05:31.766 ************************************ 00:05:31.766 END TEST locking_overlapped_coremask 00:05:31.766 ************************************ 00:05:31.766 17:51:25 event.cpu_locks -- common/autotest_common.sh@1142 -- # return 0 00:05:31.766 17:51:25 event.cpu_locks -- event/cpu_locks.sh@172 -- # run_test locking_overlapped_coremask_via_rpc locking_overlapped_coremask_via_rpc 00:05:31.766 17:51:25 event.cpu_locks -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:05:31.766 17:51:25 event.cpu_locks -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:31.766 17:51:25 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:05:31.766 ************************************ 00:05:31.766 START TEST locking_overlapped_coremask_via_rpc 00:05:31.766 ************************************ 00:05:31.766 17:51:25 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1123 -- # locking_overlapped_coremask_via_rpc 00:05:31.766 17:51:25 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@148 -- # spdk_tgt_pid=418818 00:05:31.766 17:51:25 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@149 -- # waitforlisten 418818 /var/tmp/spdk.sock 00:05:31.766 17:51:25 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@147 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x7 --disable-cpumask-locks 00:05:31.766 17:51:25 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@829 -- # '[' -z 418818 ']' 00:05:31.766 17:51:25 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:31.766 17:51:25 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@834 -- # local max_retries=100 00:05:31.766 17:51:25 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:31.766 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:31.766 17:51:25 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # xtrace_disable 00:05:31.766 17:51:25 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:32.025 [2024-07-15 17:51:25.497678] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:32.025 [2024-07-15 17:51:25.497718] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid418818 ] 00:05:32.025 EAL: No free 2048 kB hugepages reported on node 1 00:05:32.025 [2024-07-15 17:51:25.550379] app.c: 906:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:05:32.025 [2024-07-15 17:51:25.550401] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:05:32.025 [2024-07-15 17:51:25.631635] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:05:32.025 [2024-07-15 17:51:25.631732] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:32.025 [2024-07-15 17:51:25.631732] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:05:32.593 17:51:26 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:05:32.593 17:51:26 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@862 -- # return 0 00:05:32.593 17:51:26 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@152 -- # spdk_tgt_pid2=419014 00:05:32.593 17:51:26 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@153 -- # waitforlisten 419014 /var/tmp/spdk2.sock 00:05:32.593 17:51:26 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@151 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 0x1c -r /var/tmp/spdk2.sock --disable-cpumask-locks 00:05:32.593 17:51:26 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@829 -- # '[' -z 419014 ']' 00:05:32.593 17:51:26 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk2.sock 00:05:32.593 17:51:26 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@834 -- # local max_retries=100 00:05:32.593 17:51:26 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:05:32.593 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:05:32.593 17:51:26 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # xtrace_disable 00:05:32.593 17:51:26 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:32.853 [2024-07-15 17:51:26.355125] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:32.853 [2024-07-15 17:51:26.355178] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1c --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid419014 ] 00:05:32.853 EAL: No free 2048 kB hugepages reported on node 1 00:05:32.853 [2024-07-15 17:51:26.432402] app.c: 906:spdk_app_start: *NOTICE*: CPU core locks deactivated. 00:05:32.853 [2024-07-15 17:51:26.432429] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:05:33.112 [2024-07-15 17:51:26.583475] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:05:33.112 [2024-07-15 17:51:26.583588] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:05:33.112 [2024-07-15 17:51:26.583590] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 4 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@862 -- # return 0 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@155 -- # rpc_cmd framework_enable_cpumask_locks 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@156 -- # NOT rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@648 -- # local es=0 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@650 -- # valid_exec_arg rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@636 -- # local arg=rpc_cmd 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@640 -- # type -t rpc_cmd 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@651 -- # rpc_cmd -s /var/tmp/spdk2.sock framework_enable_cpumask_locks 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:33.681 [2024-07-15 17:51:27.167299] app.c: 771:claim_cpu_cores: *ERROR*: Cannot create lock on core 2, probably process 418818 has claimed it. 00:05:33.681 request: 00:05:33.681 { 00:05:33.681 "method": "framework_enable_cpumask_locks", 00:05:33.681 "req_id": 1 00:05:33.681 } 00:05:33.681 Got JSON-RPC error response 00:05:33.681 response: 00:05:33.681 { 00:05:33.681 "code": -32603, 00:05:33.681 "message": "Failed to claim CPU core: 2" 00:05:33.681 } 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@587 -- # [[ 1 == 0 ]] 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@651 -- # es=1 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@158 -- # waitforlisten 418818 /var/tmp/spdk.sock 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@829 -- # '[' -z 418818 ']' 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@834 -- # local max_retries=100 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:33.681 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # xtrace_disable 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@862 -- # return 0 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@159 -- # waitforlisten 419014 /var/tmp/spdk2.sock 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@829 -- # '[' -z 419014 ']' 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk2.sock 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@834 -- # local max_retries=100 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock...' 00:05:33.681 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk2.sock... 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@838 -- # xtrace_disable 00:05:33.681 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:33.941 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:05:33.941 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@862 -- # return 0 00:05:33.941 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@161 -- # check_remaining_locks 00:05:33.941 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@36 -- # locks=(/var/tmp/spdk_cpu_lock_*) 00:05:33.941 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@37 -- # locks_expected=(/var/tmp/spdk_cpu_lock_{000..002}) 00:05:33.941 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- event/cpu_locks.sh@38 -- # [[ /var/tmp/spdk_cpu_lock_000 /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 == \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\0\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\1\ \/\v\a\r\/\t\m\p\/\s\p\d\k\_\c\p\u\_\l\o\c\k\_\0\0\2 ]] 00:05:33.941 00:05:33.941 real 0m2.095s 00:05:33.941 user 0m0.871s 00:05:33.941 sys 0m0.151s 00:05:33.941 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:33.941 17:51:27 event.cpu_locks.locking_overlapped_coremask_via_rpc -- common/autotest_common.sh@10 -- # set +x 00:05:33.941 ************************************ 00:05:33.941 END TEST locking_overlapped_coremask_via_rpc 00:05:33.941 ************************************ 00:05:33.941 17:51:27 event.cpu_locks -- common/autotest_common.sh@1142 -- # return 0 00:05:33.941 17:51:27 event.cpu_locks -- event/cpu_locks.sh@174 -- # cleanup 00:05:33.941 17:51:27 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 418818 ]] 00:05:33.941 17:51:27 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 418818 00:05:33.941 17:51:27 event.cpu_locks -- common/autotest_common.sh@948 -- # '[' -z 418818 ']' 00:05:33.941 17:51:27 event.cpu_locks -- common/autotest_common.sh@952 -- # kill -0 418818 00:05:33.941 17:51:27 event.cpu_locks -- common/autotest_common.sh@953 -- # uname 00:05:33.941 17:51:27 event.cpu_locks -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:05:33.941 17:51:27 event.cpu_locks -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 418818 00:05:33.941 17:51:27 event.cpu_locks -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:05:33.941 17:51:27 event.cpu_locks -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:05:33.941 17:51:27 event.cpu_locks -- common/autotest_common.sh@966 -- # echo 'killing process with pid 418818' 00:05:33.941 killing process with pid 418818 00:05:33.941 17:51:27 event.cpu_locks -- common/autotest_common.sh@967 -- # kill 418818 00:05:33.941 17:51:27 event.cpu_locks -- common/autotest_common.sh@972 -- # wait 418818 00:05:34.510 17:51:27 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 419014 ]] 00:05:34.510 17:51:27 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 419014 00:05:34.510 17:51:27 event.cpu_locks -- common/autotest_common.sh@948 -- # '[' -z 419014 ']' 00:05:34.510 17:51:27 event.cpu_locks -- common/autotest_common.sh@952 -- # kill -0 419014 00:05:34.510 17:51:27 event.cpu_locks -- common/autotest_common.sh@953 -- # uname 00:05:34.510 17:51:27 event.cpu_locks -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:05:34.510 17:51:27 event.cpu_locks -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 419014 00:05:34.510 17:51:27 event.cpu_locks -- common/autotest_common.sh@954 -- # process_name=reactor_2 00:05:34.510 17:51:27 event.cpu_locks -- common/autotest_common.sh@958 -- # '[' reactor_2 = sudo ']' 00:05:34.510 17:51:27 event.cpu_locks -- common/autotest_common.sh@966 -- # echo 'killing process with pid 419014' 00:05:34.510 killing process with pid 419014 00:05:34.510 17:51:27 event.cpu_locks -- common/autotest_common.sh@967 -- # kill 419014 00:05:34.510 17:51:27 event.cpu_locks -- common/autotest_common.sh@972 -- # wait 419014 00:05:34.769 17:51:28 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:05:34.769 17:51:28 event.cpu_locks -- event/cpu_locks.sh@1 -- # cleanup 00:05:34.769 17:51:28 event.cpu_locks -- event/cpu_locks.sh@15 -- # [[ -z 418818 ]] 00:05:34.769 17:51:28 event.cpu_locks -- event/cpu_locks.sh@15 -- # killprocess 418818 00:05:34.769 17:51:28 event.cpu_locks -- common/autotest_common.sh@948 -- # '[' -z 418818 ']' 00:05:34.769 17:51:28 event.cpu_locks -- common/autotest_common.sh@952 -- # kill -0 418818 00:05:34.769 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 952: kill: (418818) - No such process 00:05:34.769 17:51:28 event.cpu_locks -- common/autotest_common.sh@975 -- # echo 'Process with pid 418818 is not found' 00:05:34.769 Process with pid 418818 is not found 00:05:34.769 17:51:28 event.cpu_locks -- event/cpu_locks.sh@16 -- # [[ -z 419014 ]] 00:05:34.769 17:51:28 event.cpu_locks -- event/cpu_locks.sh@16 -- # killprocess 419014 00:05:34.769 17:51:28 event.cpu_locks -- common/autotest_common.sh@948 -- # '[' -z 419014 ']' 00:05:34.769 17:51:28 event.cpu_locks -- common/autotest_common.sh@952 -- # kill -0 419014 00:05:34.769 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 952: kill: (419014) - No such process 00:05:34.769 17:51:28 event.cpu_locks -- common/autotest_common.sh@975 -- # echo 'Process with pid 419014 is not found' 00:05:34.769 Process with pid 419014 is not found 00:05:34.769 17:51:28 event.cpu_locks -- event/cpu_locks.sh@18 -- # rm -f 00:05:34.769 00:05:34.769 real 0m16.580s 00:05:34.769 user 0m28.827s 00:05:34.798 sys 0m4.626s 00:05:34.798 17:51:28 event.cpu_locks -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:34.798 17:51:28 event.cpu_locks -- common/autotest_common.sh@10 -- # set +x 00:05:34.798 ************************************ 00:05:34.798 END TEST cpu_locks 00:05:34.798 ************************************ 00:05:34.798 17:51:28 event -- common/autotest_common.sh@1142 -- # return 0 00:05:34.798 00:05:34.798 real 0m41.148s 00:05:34.798 user 1m18.692s 00:05:34.798 sys 0m7.866s 00:05:34.798 17:51:28 event -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:34.798 17:51:28 event -- common/autotest_common.sh@10 -- # set +x 00:05:34.798 ************************************ 00:05:34.798 END TEST event 00:05:34.798 ************************************ 00:05:34.798 17:51:28 -- common/autotest_common.sh@1142 -- # return 0 00:05:34.798 17:51:28 -- spdk/autotest.sh@182 -- # run_test thread /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/thread.sh 00:05:34.798 17:51:28 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:05:34.798 17:51:28 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:34.798 17:51:28 -- common/autotest_common.sh@10 -- # set +x 00:05:34.798 ************************************ 00:05:34.798 START TEST thread 00:05:34.798 ************************************ 00:05:34.798 17:51:28 thread -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/thread.sh 00:05:34.798 * Looking for test storage... 00:05:35.057 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread 00:05:35.057 17:51:28 thread -- thread/thread.sh@11 -- # run_test thread_poller_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:05:35.057 17:51:28 thread -- common/autotest_common.sh@1099 -- # '[' 8 -le 1 ']' 00:05:35.057 17:51:28 thread -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:35.057 17:51:28 thread -- common/autotest_common.sh@10 -- # set +x 00:05:35.057 ************************************ 00:05:35.057 START TEST thread_poller_perf 00:05:35.057 ************************************ 00:05:35.057 17:51:28 thread.thread_poller_perf -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 1 -t 1 00:05:35.057 [2024-07-15 17:51:28.554459] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:35.057 [2024-07-15 17:51:28.554529] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid419389 ] 00:05:35.057 EAL: No free 2048 kB hugepages reported on node 1 00:05:35.057 [2024-07-15 17:51:28.608545] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:35.057 [2024-07-15 17:51:28.681404] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:35.057 Running 1000 pollers for 1 seconds with 1 microseconds period. 00:05:36.433 ====================================== 00:05:36.433 busy:2305759186 (cyc) 00:05:36.433 total_run_count: 412000 00:05:36.433 tsc_hz: 2300000000 (cyc) 00:05:36.433 ====================================== 00:05:36.433 poller_cost: 5596 (cyc), 2433 (nsec) 00:05:36.433 00:05:36.433 real 0m1.221s 00:05:36.433 user 0m1.146s 00:05:36.433 sys 0m0.069s 00:05:36.433 17:51:29 thread.thread_poller_perf -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:36.433 17:51:29 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:05:36.433 ************************************ 00:05:36.433 END TEST thread_poller_perf 00:05:36.433 ************************************ 00:05:36.433 17:51:29 thread -- common/autotest_common.sh@1142 -- # return 0 00:05:36.433 17:51:29 thread -- thread/thread.sh@12 -- # run_test thread_poller_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:05:36.433 17:51:29 thread -- common/autotest_common.sh@1099 -- # '[' 8 -le 1 ']' 00:05:36.433 17:51:29 thread -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:36.433 17:51:29 thread -- common/autotest_common.sh@10 -- # set +x 00:05:36.433 ************************************ 00:05:36.433 START TEST thread_poller_perf 00:05:36.433 ************************************ 00:05:36.433 17:51:29 thread.thread_poller_perf -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/thread/poller_perf/poller_perf -b 1000 -l 0 -t 1 00:05:36.433 [2024-07-15 17:51:29.834628] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:36.433 [2024-07-15 17:51:29.834695] [ DPDK EAL parameters: poller_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid419641 ] 00:05:36.433 EAL: No free 2048 kB hugepages reported on node 1 00:05:36.433 [2024-07-15 17:51:29.891089] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:36.433 [2024-07-15 17:51:29.962131] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:36.433 Running 1000 pollers for 1 seconds with 0 microseconds period. 00:05:37.383 ====================================== 00:05:37.383 busy:2301459818 (cyc) 00:05:37.383 total_run_count: 5236000 00:05:37.383 tsc_hz: 2300000000 (cyc) 00:05:37.383 ====================================== 00:05:37.383 poller_cost: 439 (cyc), 190 (nsec) 00:05:37.383 00:05:37.383 real 0m1.219s 00:05:37.383 user 0m1.149s 00:05:37.383 sys 0m0.066s 00:05:37.383 17:51:31 thread.thread_poller_perf -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:37.383 17:51:31 thread.thread_poller_perf -- common/autotest_common.sh@10 -- # set +x 00:05:37.383 ************************************ 00:05:37.383 END TEST thread_poller_perf 00:05:37.383 ************************************ 00:05:37.383 17:51:31 thread -- common/autotest_common.sh@1142 -- # return 0 00:05:37.383 17:51:31 thread -- thread/thread.sh@17 -- # [[ y != \y ]] 00:05:37.383 00:05:37.383 real 0m2.647s 00:05:37.383 user 0m2.379s 00:05:37.383 sys 0m0.273s 00:05:37.383 17:51:31 thread -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:37.383 17:51:31 thread -- common/autotest_common.sh@10 -- # set +x 00:05:37.383 ************************************ 00:05:37.383 END TEST thread 00:05:37.383 ************************************ 00:05:37.383 17:51:31 -- common/autotest_common.sh@1142 -- # return 0 00:05:37.383 17:51:31 -- spdk/autotest.sh@183 -- # run_test accel /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/accel.sh 00:05:37.383 17:51:31 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:05:37.383 17:51:31 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:37.383 17:51:31 -- common/autotest_common.sh@10 -- # set +x 00:05:37.641 ************************************ 00:05:37.641 START TEST accel 00:05:37.641 ************************************ 00:05:37.641 17:51:31 accel -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/accel.sh 00:05:37.641 * Looking for test storage... 00:05:37.641 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel 00:05:37.641 17:51:31 accel -- accel/accel.sh@81 -- # declare -A expected_opcs 00:05:37.641 17:51:31 accel -- accel/accel.sh@82 -- # get_expected_opcs 00:05:37.641 17:51:31 accel -- accel/accel.sh@60 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:05:37.641 17:51:31 accel -- accel/accel.sh@62 -- # spdk_tgt_pid=419929 00:05:37.641 17:51:31 accel -- accel/accel.sh@63 -- # waitforlisten 419929 00:05:37.641 17:51:31 accel -- common/autotest_common.sh@829 -- # '[' -z 419929 ']' 00:05:37.641 17:51:31 accel -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:05:37.641 17:51:31 accel -- accel/accel.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -c /dev/fd/63 00:05:37.641 17:51:31 accel -- common/autotest_common.sh@834 -- # local max_retries=100 00:05:37.641 17:51:31 accel -- accel/accel.sh@61 -- # build_accel_config 00:05:37.641 17:51:31 accel -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:05:37.641 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:05:37.641 17:51:31 accel -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:37.641 17:51:31 accel -- common/autotest_common.sh@838 -- # xtrace_disable 00:05:37.641 17:51:31 accel -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:37.641 17:51:31 accel -- common/autotest_common.sh@10 -- # set +x 00:05:37.641 17:51:31 accel -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:37.641 17:51:31 accel -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:37.641 17:51:31 accel -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:37.641 17:51:31 accel -- accel/accel.sh@40 -- # local IFS=, 00:05:37.641 17:51:31 accel -- accel/accel.sh@41 -- # jq -r . 00:05:37.641 [2024-07-15 17:51:31.255618] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:37.641 [2024-07-15 17:51:31.255666] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid419929 ] 00:05:37.641 EAL: No free 2048 kB hugepages reported on node 1 00:05:37.641 [2024-07-15 17:51:31.309335] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:37.900 [2024-07-15 17:51:31.383145] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:38.467 17:51:32 accel -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:05:38.467 17:51:32 accel -- common/autotest_common.sh@862 -- # return 0 00:05:38.467 17:51:32 accel -- accel/accel.sh@65 -- # [[ 0 -gt 0 ]] 00:05:38.467 17:51:32 accel -- accel/accel.sh@66 -- # [[ 0 -gt 0 ]] 00:05:38.467 17:51:32 accel -- accel/accel.sh@67 -- # [[ 0 -gt 0 ]] 00:05:38.467 17:51:32 accel -- accel/accel.sh@68 -- # [[ -n '' ]] 00:05:38.467 17:51:32 accel -- accel/accel.sh@70 -- # exp_opcs=($($rpc_py accel_get_opc_assignments | jq -r ". | to_entries | map(\"\(.key)=\(.value)\") | .[]")) 00:05:38.467 17:51:32 accel -- accel/accel.sh@70 -- # rpc_cmd accel_get_opc_assignments 00:05:38.467 17:51:32 accel -- accel/accel.sh@70 -- # jq -r '. | to_entries | map("\(.key)=\(.value)") | .[]' 00:05:38.467 17:51:32 accel -- common/autotest_common.sh@559 -- # xtrace_disable 00:05:38.467 17:51:32 accel -- common/autotest_common.sh@10 -- # set +x 00:05:38.467 17:51:32 accel -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:05:38.467 17:51:32 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # IFS== 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # read -r opc module 00:05:38.467 17:51:32 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:05:38.467 17:51:32 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # IFS== 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # read -r opc module 00:05:38.467 17:51:32 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:05:38.467 17:51:32 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # IFS== 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # read -r opc module 00:05:38.467 17:51:32 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:05:38.467 17:51:32 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # IFS== 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # read -r opc module 00:05:38.467 17:51:32 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:05:38.467 17:51:32 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # IFS== 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # read -r opc module 00:05:38.467 17:51:32 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:05:38.467 17:51:32 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # IFS== 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # read -r opc module 00:05:38.467 17:51:32 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:05:38.467 17:51:32 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # IFS== 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # read -r opc module 00:05:38.467 17:51:32 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:05:38.467 17:51:32 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # IFS== 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # read -r opc module 00:05:38.467 17:51:32 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:05:38.467 17:51:32 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # IFS== 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # read -r opc module 00:05:38.467 17:51:32 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:05:38.467 17:51:32 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # IFS== 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # read -r opc module 00:05:38.467 17:51:32 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:05:38.467 17:51:32 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # IFS== 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # read -r opc module 00:05:38.467 17:51:32 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:05:38.467 17:51:32 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # IFS== 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # read -r opc module 00:05:38.467 17:51:32 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:05:38.467 17:51:32 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # IFS== 00:05:38.467 17:51:32 accel -- accel/accel.sh@72 -- # read -r opc module 00:05:38.467 17:51:32 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:05:38.468 17:51:32 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:05:38.468 17:51:32 accel -- accel/accel.sh@72 -- # IFS== 00:05:38.468 17:51:32 accel -- accel/accel.sh@72 -- # read -r opc module 00:05:38.468 17:51:32 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:05:38.468 17:51:32 accel -- accel/accel.sh@71 -- # for opc_opt in "${exp_opcs[@]}" 00:05:38.468 17:51:32 accel -- accel/accel.sh@72 -- # IFS== 00:05:38.468 17:51:32 accel -- accel/accel.sh@72 -- # read -r opc module 00:05:38.468 17:51:32 accel -- accel/accel.sh@73 -- # expected_opcs["$opc"]=software 00:05:38.468 17:51:32 accel -- accel/accel.sh@75 -- # killprocess 419929 00:05:38.468 17:51:32 accel -- common/autotest_common.sh@948 -- # '[' -z 419929 ']' 00:05:38.468 17:51:32 accel -- common/autotest_common.sh@952 -- # kill -0 419929 00:05:38.468 17:51:32 accel -- common/autotest_common.sh@953 -- # uname 00:05:38.468 17:51:32 accel -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:05:38.468 17:51:32 accel -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 419929 00:05:38.468 17:51:32 accel -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:05:38.468 17:51:32 accel -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:05:38.468 17:51:32 accel -- common/autotest_common.sh@966 -- # echo 'killing process with pid 419929' 00:05:38.468 killing process with pid 419929 00:05:38.468 17:51:32 accel -- common/autotest_common.sh@967 -- # kill 419929 00:05:38.468 17:51:32 accel -- common/autotest_common.sh@972 -- # wait 419929 00:05:38.726 17:51:32 accel -- accel/accel.sh@76 -- # trap - ERR 00:05:38.726 17:51:32 accel -- accel/accel.sh@89 -- # run_test accel_help accel_perf -h 00:05:38.986 17:51:32 accel -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:05:38.986 17:51:32 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:38.986 17:51:32 accel -- common/autotest_common.sh@10 -- # set +x 00:05:38.986 17:51:32 accel.accel_help -- common/autotest_common.sh@1123 -- # accel_perf -h 00:05:38.986 17:51:32 accel.accel_help -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -h 00:05:38.986 17:51:32 accel.accel_help -- accel/accel.sh@12 -- # build_accel_config 00:05:38.986 17:51:32 accel.accel_help -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:38.986 17:51:32 accel.accel_help -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:38.986 17:51:32 accel.accel_help -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:38.986 17:51:32 accel.accel_help -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:38.986 17:51:32 accel.accel_help -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:38.986 17:51:32 accel.accel_help -- accel/accel.sh@40 -- # local IFS=, 00:05:38.986 17:51:32 accel.accel_help -- accel/accel.sh@41 -- # jq -r . 00:05:38.986 17:51:32 accel.accel_help -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:38.986 17:51:32 accel.accel_help -- common/autotest_common.sh@10 -- # set +x 00:05:38.986 17:51:32 accel -- common/autotest_common.sh@1142 -- # return 0 00:05:38.986 17:51:32 accel -- accel/accel.sh@91 -- # run_test accel_missing_filename NOT accel_perf -t 1 -w compress 00:05:38.986 17:51:32 accel -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:05:38.986 17:51:32 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:38.986 17:51:32 accel -- common/autotest_common.sh@10 -- # set +x 00:05:38.986 ************************************ 00:05:38.986 START TEST accel_missing_filename 00:05:38.986 ************************************ 00:05:38.986 17:51:32 accel.accel_missing_filename -- common/autotest_common.sh@1123 -- # NOT accel_perf -t 1 -w compress 00:05:38.986 17:51:32 accel.accel_missing_filename -- common/autotest_common.sh@648 -- # local es=0 00:05:38.986 17:51:32 accel.accel_missing_filename -- common/autotest_common.sh@650 -- # valid_exec_arg accel_perf -t 1 -w compress 00:05:38.986 17:51:32 accel.accel_missing_filename -- common/autotest_common.sh@636 -- # local arg=accel_perf 00:05:38.986 17:51:32 accel.accel_missing_filename -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:05:38.986 17:51:32 accel.accel_missing_filename -- common/autotest_common.sh@640 -- # type -t accel_perf 00:05:38.986 17:51:32 accel.accel_missing_filename -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:05:38.986 17:51:32 accel.accel_missing_filename -- common/autotest_common.sh@651 -- # accel_perf -t 1 -w compress 00:05:38.986 17:51:32 accel.accel_missing_filename -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w compress 00:05:38.986 17:51:32 accel.accel_missing_filename -- accel/accel.sh@12 -- # build_accel_config 00:05:38.986 17:51:32 accel.accel_missing_filename -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:38.986 17:51:32 accel.accel_missing_filename -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:38.986 17:51:32 accel.accel_missing_filename -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:38.986 17:51:32 accel.accel_missing_filename -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:38.986 17:51:32 accel.accel_missing_filename -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:38.986 17:51:32 accel.accel_missing_filename -- accel/accel.sh@40 -- # local IFS=, 00:05:38.986 17:51:32 accel.accel_missing_filename -- accel/accel.sh@41 -- # jq -r . 00:05:38.986 [2024-07-15 17:51:32.606915] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:38.986 [2024-07-15 17:51:32.606980] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid420193 ] 00:05:38.986 EAL: No free 2048 kB hugepages reported on node 1 00:05:38.986 [2024-07-15 17:51:32.664442] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:39.245 [2024-07-15 17:51:32.736880] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:39.245 [2024-07-15 17:51:32.777969] app.c:1053:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:05:39.245 [2024-07-15 17:51:32.837821] accel_perf.c:1463:main: *ERROR*: ERROR starting application 00:05:39.245 A filename is required. 00:05:39.245 17:51:32 accel.accel_missing_filename -- common/autotest_common.sh@651 -- # es=234 00:05:39.245 17:51:32 accel.accel_missing_filename -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:05:39.245 17:51:32 accel.accel_missing_filename -- common/autotest_common.sh@660 -- # es=106 00:05:39.245 17:51:32 accel.accel_missing_filename -- common/autotest_common.sh@661 -- # case "$es" in 00:05:39.245 17:51:32 accel.accel_missing_filename -- common/autotest_common.sh@668 -- # es=1 00:05:39.245 17:51:32 accel.accel_missing_filename -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:05:39.245 00:05:39.245 real 0m0.335s 00:05:39.245 user 0m0.254s 00:05:39.245 sys 0m0.118s 00:05:39.245 17:51:32 accel.accel_missing_filename -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:39.245 17:51:32 accel.accel_missing_filename -- common/autotest_common.sh@10 -- # set +x 00:05:39.245 ************************************ 00:05:39.245 END TEST accel_missing_filename 00:05:39.245 ************************************ 00:05:39.245 17:51:32 accel -- common/autotest_common.sh@1142 -- # return 0 00:05:39.245 17:51:32 accel -- accel/accel.sh@93 -- # run_test accel_compress_verify NOT accel_perf -t 1 -w compress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y 00:05:39.245 17:51:32 accel -- common/autotest_common.sh@1099 -- # '[' 10 -le 1 ']' 00:05:39.245 17:51:32 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:39.245 17:51:32 accel -- common/autotest_common.sh@10 -- # set +x 00:05:39.245 ************************************ 00:05:39.245 START TEST accel_compress_verify 00:05:39.245 ************************************ 00:05:39.245 17:51:32 accel.accel_compress_verify -- common/autotest_common.sh@1123 -- # NOT accel_perf -t 1 -w compress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y 00:05:39.246 17:51:32 accel.accel_compress_verify -- common/autotest_common.sh@648 -- # local es=0 00:05:39.246 17:51:32 accel.accel_compress_verify -- common/autotest_common.sh@650 -- # valid_exec_arg accel_perf -t 1 -w compress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y 00:05:39.246 17:51:32 accel.accel_compress_verify -- common/autotest_common.sh@636 -- # local arg=accel_perf 00:05:39.505 17:51:32 accel.accel_compress_verify -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:05:39.505 17:51:32 accel.accel_compress_verify -- common/autotest_common.sh@640 -- # type -t accel_perf 00:05:39.505 17:51:32 accel.accel_compress_verify -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:05:39.505 17:51:32 accel.accel_compress_verify -- common/autotest_common.sh@651 -- # accel_perf -t 1 -w compress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y 00:05:39.505 17:51:32 accel.accel_compress_verify -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w compress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y 00:05:39.505 17:51:32 accel.accel_compress_verify -- accel/accel.sh@12 -- # build_accel_config 00:05:39.505 17:51:32 accel.accel_compress_verify -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:39.505 17:51:32 accel.accel_compress_verify -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:39.505 17:51:32 accel.accel_compress_verify -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:39.505 17:51:32 accel.accel_compress_verify -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:39.505 17:51:32 accel.accel_compress_verify -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:39.505 17:51:32 accel.accel_compress_verify -- accel/accel.sh@40 -- # local IFS=, 00:05:39.505 17:51:32 accel.accel_compress_verify -- accel/accel.sh@41 -- # jq -r . 00:05:39.505 [2024-07-15 17:51:33.001455] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:39.505 [2024-07-15 17:51:33.001522] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid420299 ] 00:05:39.505 EAL: No free 2048 kB hugepages reported on node 1 00:05:39.505 [2024-07-15 17:51:33.059437] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:39.505 [2024-07-15 17:51:33.132500] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:39.505 [2024-07-15 17:51:33.173877] app.c:1053:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:05:39.764 [2024-07-15 17:51:33.233951] accel_perf.c:1463:main: *ERROR*: ERROR starting application 00:05:39.764 00:05:39.764 Compression does not support the verify option, aborting. 00:05:39.764 17:51:33 accel.accel_compress_verify -- common/autotest_common.sh@651 -- # es=161 00:05:39.764 17:51:33 accel.accel_compress_verify -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:05:39.764 17:51:33 accel.accel_compress_verify -- common/autotest_common.sh@660 -- # es=33 00:05:39.764 17:51:33 accel.accel_compress_verify -- common/autotest_common.sh@661 -- # case "$es" in 00:05:39.764 17:51:33 accel.accel_compress_verify -- common/autotest_common.sh@668 -- # es=1 00:05:39.764 17:51:33 accel.accel_compress_verify -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:05:39.764 00:05:39.764 real 0m0.335s 00:05:39.764 user 0m0.257s 00:05:39.764 sys 0m0.120s 00:05:39.764 17:51:33 accel.accel_compress_verify -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:39.764 17:51:33 accel.accel_compress_verify -- common/autotest_common.sh@10 -- # set +x 00:05:39.764 ************************************ 00:05:39.764 END TEST accel_compress_verify 00:05:39.764 ************************************ 00:05:39.764 17:51:33 accel -- common/autotest_common.sh@1142 -- # return 0 00:05:39.764 17:51:33 accel -- accel/accel.sh@95 -- # run_test accel_wrong_workload NOT accel_perf -t 1 -w foobar 00:05:39.764 17:51:33 accel -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:05:39.764 17:51:33 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:39.764 17:51:33 accel -- common/autotest_common.sh@10 -- # set +x 00:05:39.764 ************************************ 00:05:39.764 START TEST accel_wrong_workload 00:05:39.764 ************************************ 00:05:39.764 17:51:33 accel.accel_wrong_workload -- common/autotest_common.sh@1123 -- # NOT accel_perf -t 1 -w foobar 00:05:39.764 17:51:33 accel.accel_wrong_workload -- common/autotest_common.sh@648 -- # local es=0 00:05:39.764 17:51:33 accel.accel_wrong_workload -- common/autotest_common.sh@650 -- # valid_exec_arg accel_perf -t 1 -w foobar 00:05:39.764 17:51:33 accel.accel_wrong_workload -- common/autotest_common.sh@636 -- # local arg=accel_perf 00:05:39.764 17:51:33 accel.accel_wrong_workload -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:05:39.764 17:51:33 accel.accel_wrong_workload -- common/autotest_common.sh@640 -- # type -t accel_perf 00:05:39.764 17:51:33 accel.accel_wrong_workload -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:05:39.764 17:51:33 accel.accel_wrong_workload -- common/autotest_common.sh@651 -- # accel_perf -t 1 -w foobar 00:05:39.764 17:51:33 accel.accel_wrong_workload -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w foobar 00:05:39.764 17:51:33 accel.accel_wrong_workload -- accel/accel.sh@12 -- # build_accel_config 00:05:39.764 17:51:33 accel.accel_wrong_workload -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:39.764 17:51:33 accel.accel_wrong_workload -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:39.764 17:51:33 accel.accel_wrong_workload -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:39.764 17:51:33 accel.accel_wrong_workload -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:39.764 17:51:33 accel.accel_wrong_workload -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:39.764 17:51:33 accel.accel_wrong_workload -- accel/accel.sh@40 -- # local IFS=, 00:05:39.764 17:51:33 accel.accel_wrong_workload -- accel/accel.sh@41 -- # jq -r . 00:05:39.764 Unsupported workload type: foobar 00:05:39.764 [2024-07-15 17:51:33.389118] app.c:1451:spdk_app_parse_args: *ERROR*: Parsing app-specific command line parameter 'w' failed: 1 00:05:39.764 accel_perf options: 00:05:39.764 [-h help message] 00:05:39.764 [-q queue depth per core] 00:05:39.764 [-C for supported workloads, use this value to configure the io vector size to test (default 1) 00:05:39.764 [-T number of threads per core 00:05:39.764 [-o transfer size in bytes (default: 4KiB. For compress/decompress, 0 means the input file size)] 00:05:39.764 [-t time in seconds] 00:05:39.764 [-w workload type must be one of these: copy, fill, crc32c, copy_crc32c, compare, compress, decompress, dualcast, xor, 00:05:39.764 [ dif_verify, dif_verify_copy, dif_generate, dif_generate_copy 00:05:39.764 [-M assign module to the operation, not compatible with accel_assign_opc RPC 00:05:39.764 [-l for compress/decompress workloads, name of uncompressed input file 00:05:39.764 [-S for crc32c workload, use this seed value (default 0) 00:05:39.764 [-P for compare workload, percentage of operations that should miscompare (percent, default 0) 00:05:39.764 [-f for fill workload, use this BYTE value (default 255) 00:05:39.764 [-x for xor workload, use this number of source buffers (default, minimum: 2)] 00:05:39.764 [-y verify result if this switch is on] 00:05:39.764 [-a tasks to allocate per core (default: same value as -q)] 00:05:39.764 Can be used to spread operations across a wider range of memory. 00:05:39.764 17:51:33 accel.accel_wrong_workload -- common/autotest_common.sh@651 -- # es=1 00:05:39.764 17:51:33 accel.accel_wrong_workload -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:05:39.764 17:51:33 accel.accel_wrong_workload -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:05:39.764 17:51:33 accel.accel_wrong_workload -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:05:39.764 00:05:39.764 real 0m0.030s 00:05:39.764 user 0m0.016s 00:05:39.764 sys 0m0.014s 00:05:39.764 17:51:33 accel.accel_wrong_workload -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:39.764 17:51:33 accel.accel_wrong_workload -- common/autotest_common.sh@10 -- # set +x 00:05:39.764 ************************************ 00:05:39.764 END TEST accel_wrong_workload 00:05:39.764 ************************************ 00:05:39.764 Error: writing output failed: Broken pipe 00:05:39.764 17:51:33 accel -- common/autotest_common.sh@1142 -- # return 0 00:05:39.764 17:51:33 accel -- accel/accel.sh@97 -- # run_test accel_negative_buffers NOT accel_perf -t 1 -w xor -y -x -1 00:05:39.764 17:51:33 accel -- common/autotest_common.sh@1099 -- # '[' 10 -le 1 ']' 00:05:39.764 17:51:33 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:39.764 17:51:33 accel -- common/autotest_common.sh@10 -- # set +x 00:05:39.764 ************************************ 00:05:39.764 START TEST accel_negative_buffers 00:05:39.764 ************************************ 00:05:39.764 17:51:33 accel.accel_negative_buffers -- common/autotest_common.sh@1123 -- # NOT accel_perf -t 1 -w xor -y -x -1 00:05:39.764 17:51:33 accel.accel_negative_buffers -- common/autotest_common.sh@648 -- # local es=0 00:05:39.764 17:51:33 accel.accel_negative_buffers -- common/autotest_common.sh@650 -- # valid_exec_arg accel_perf -t 1 -w xor -y -x -1 00:05:39.764 17:51:33 accel.accel_negative_buffers -- common/autotest_common.sh@636 -- # local arg=accel_perf 00:05:39.764 17:51:33 accel.accel_negative_buffers -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:05:39.764 17:51:33 accel.accel_negative_buffers -- common/autotest_common.sh@640 -- # type -t accel_perf 00:05:39.764 17:51:33 accel.accel_negative_buffers -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:05:39.764 17:51:33 accel.accel_negative_buffers -- common/autotest_common.sh@651 -- # accel_perf -t 1 -w xor -y -x -1 00:05:39.764 17:51:33 accel.accel_negative_buffers -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w xor -y -x -1 00:05:39.764 17:51:33 accel.accel_negative_buffers -- accel/accel.sh@12 -- # build_accel_config 00:05:39.764 17:51:33 accel.accel_negative_buffers -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:39.764 17:51:33 accel.accel_negative_buffers -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:39.764 17:51:33 accel.accel_negative_buffers -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:39.764 17:51:33 accel.accel_negative_buffers -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:39.764 17:51:33 accel.accel_negative_buffers -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:39.764 17:51:33 accel.accel_negative_buffers -- accel/accel.sh@40 -- # local IFS=, 00:05:39.764 17:51:33 accel.accel_negative_buffers -- accel/accel.sh@41 -- # jq -r . 00:05:39.764 -x option must be non-negative. 00:05:39.764 [2024-07-15 17:51:33.478035] app.c:1451:spdk_app_parse_args: *ERROR*: Parsing app-specific command line parameter 'x' failed: 1 00:05:39.764 accel_perf options: 00:05:39.764 [-h help message] 00:05:39.764 [-q queue depth per core] 00:05:39.764 [-C for supported workloads, use this value to configure the io vector size to test (default 1) 00:05:39.764 [-T number of threads per core 00:05:39.764 [-o transfer size in bytes (default: 4KiB. For compress/decompress, 0 means the input file size)] 00:05:39.764 [-t time in seconds] 00:05:39.764 [-w workload type must be one of these: copy, fill, crc32c, copy_crc32c, compare, compress, decompress, dualcast, xor, 00:05:39.764 [ dif_verify, dif_verify_copy, dif_generate, dif_generate_copy 00:05:39.764 [-M assign module to the operation, not compatible with accel_assign_opc RPC 00:05:39.764 [-l for compress/decompress workloads, name of uncompressed input file 00:05:39.764 [-S for crc32c workload, use this seed value (default 0) 00:05:39.764 [-P for compare workload, percentage of operations that should miscompare (percent, default 0) 00:05:39.764 [-f for fill workload, use this BYTE value (default 255) 00:05:39.764 [-x for xor workload, use this number of source buffers (default, minimum: 2)] 00:05:39.764 [-y verify result if this switch is on] 00:05:39.764 [-a tasks to allocate per core (default: same value as -q)] 00:05:39.764 Can be used to spread operations across a wider range of memory. 00:05:39.764 17:51:33 accel.accel_negative_buffers -- common/autotest_common.sh@651 -- # es=1 00:05:39.764 17:51:33 accel.accel_negative_buffers -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:05:39.764 17:51:33 accel.accel_negative_buffers -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:05:39.764 17:51:33 accel.accel_negative_buffers -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:05:39.764 00:05:39.764 real 0m0.031s 00:05:39.764 user 0m0.021s 00:05:39.764 sys 0m0.010s 00:05:39.764 17:51:33 accel.accel_negative_buffers -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:39.764 17:51:33 accel.accel_negative_buffers -- common/autotest_common.sh@10 -- # set +x 00:05:39.764 ************************************ 00:05:39.764 END TEST accel_negative_buffers 00:05:39.764 ************************************ 00:05:40.023 Error: writing output failed: Broken pipe 00:05:40.023 17:51:33 accel -- common/autotest_common.sh@1142 -- # return 0 00:05:40.023 17:51:33 accel -- accel/accel.sh@101 -- # run_test accel_crc32c accel_test -t 1 -w crc32c -S 32 -y 00:05:40.023 17:51:33 accel -- common/autotest_common.sh@1099 -- # '[' 9 -le 1 ']' 00:05:40.023 17:51:33 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:40.023 17:51:33 accel -- common/autotest_common.sh@10 -- # set +x 00:05:40.023 ************************************ 00:05:40.023 START TEST accel_crc32c 00:05:40.023 ************************************ 00:05:40.023 17:51:33 accel.accel_crc32c -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w crc32c -S 32 -y 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@16 -- # local accel_opc 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@17 -- # local accel_module 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@15 -- # accel_perf -t 1 -w crc32c -S 32 -y 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w crc32c -S 32 -y 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@12 -- # build_accel_config 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@40 -- # local IFS=, 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@41 -- # jq -r . 00:05:40.023 [2024-07-15 17:51:33.563803] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:40.023 [2024-07-15 17:51:33.563851] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid420499 ] 00:05:40.023 EAL: No free 2048 kB hugepages reported on node 1 00:05:40.023 [2024-07-15 17:51:33.617729] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:40.023 [2024-07-15 17:51:33.689467] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@20 -- # val=0x1 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@20 -- # val=crc32c 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@23 -- # accel_opc=crc32c 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@20 -- # val=32 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@20 -- # val='4096 bytes' 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@20 -- # val=software 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@22 -- # accel_module=software 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@20 -- # val=32 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@20 -- # val=32 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@20 -- # val=1 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@20 -- # val='1 seconds' 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@20 -- # val=Yes 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:40.023 17:51:33 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@20 -- # val= 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@27 -- # [[ -n software ]] 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@27 -- # [[ -n crc32c ]] 00:05:41.438 17:51:34 accel.accel_crc32c -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:05:41.438 00:05:41.438 real 0m1.330s 00:05:41.438 user 0m1.229s 00:05:41.438 sys 0m0.113s 00:05:41.438 17:51:34 accel.accel_crc32c -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:41.438 17:51:34 accel.accel_crc32c -- common/autotest_common.sh@10 -- # set +x 00:05:41.438 ************************************ 00:05:41.438 END TEST accel_crc32c 00:05:41.438 ************************************ 00:05:41.438 17:51:34 accel -- common/autotest_common.sh@1142 -- # return 0 00:05:41.438 17:51:34 accel -- accel/accel.sh@102 -- # run_test accel_crc32c_C2 accel_test -t 1 -w crc32c -y -C 2 00:05:41.438 17:51:34 accel -- common/autotest_common.sh@1099 -- # '[' 9 -le 1 ']' 00:05:41.438 17:51:34 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:41.438 17:51:34 accel -- common/autotest_common.sh@10 -- # set +x 00:05:41.438 ************************************ 00:05:41.438 START TEST accel_crc32c_C2 00:05:41.438 ************************************ 00:05:41.438 17:51:34 accel.accel_crc32c_C2 -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w crc32c -y -C 2 00:05:41.438 17:51:34 accel.accel_crc32c_C2 -- accel/accel.sh@16 -- # local accel_opc 00:05:41.438 17:51:34 accel.accel_crc32c_C2 -- accel/accel.sh@17 -- # local accel_module 00:05:41.438 17:51:34 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:41.438 17:51:34 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:41.438 17:51:34 accel.accel_crc32c_C2 -- accel/accel.sh@15 -- # accel_perf -t 1 -w crc32c -y -C 2 00:05:41.438 17:51:34 accel.accel_crc32c_C2 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w crc32c -y -C 2 00:05:41.438 17:51:34 accel.accel_crc32c_C2 -- accel/accel.sh@12 -- # build_accel_config 00:05:41.438 17:51:34 accel.accel_crc32c_C2 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:41.438 17:51:34 accel.accel_crc32c_C2 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:41.438 17:51:34 accel.accel_crc32c_C2 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:41.438 17:51:34 accel.accel_crc32c_C2 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:41.438 17:51:34 accel.accel_crc32c_C2 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:41.438 17:51:34 accel.accel_crc32c_C2 -- accel/accel.sh@40 -- # local IFS=, 00:05:41.438 17:51:34 accel.accel_crc32c_C2 -- accel/accel.sh@41 -- # jq -r . 00:05:41.438 [2024-07-15 17:51:34.955710] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:41.438 [2024-07-15 17:51:34.955757] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid420755 ] 00:05:41.438 EAL: No free 2048 kB hugepages reported on node 1 00:05:41.438 [2024-07-15 17:51:35.009913] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:41.438 [2024-07-15 17:51:35.081092] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val=0x1 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val=crc32c 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@23 -- # accel_opc=crc32c 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val=0 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val='4096 bytes' 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:41.438 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val=software 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@22 -- # accel_module=software 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val=32 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val=32 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val=1 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val='1 seconds' 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val=Yes 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:41.439 17:51:35 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:42.816 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:42.816 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:42.816 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:42.816 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:42.816 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:42.816 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:42.816 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:42.816 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:42.816 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:42.816 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:42.816 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:42.816 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:42.816 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:42.816 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:42.816 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:42.817 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:42.817 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:42.817 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:42.817 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:42.817 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:42.817 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:42.817 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:42.817 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:42.817 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:42.817 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@27 -- # [[ -n software ]] 00:05:42.817 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@27 -- # [[ -n crc32c ]] 00:05:42.817 17:51:36 accel.accel_crc32c_C2 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:05:42.817 00:05:42.817 real 0m1.331s 00:05:42.817 user 0m1.232s 00:05:42.817 sys 0m0.113s 00:05:42.817 17:51:36 accel.accel_crc32c_C2 -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:42.817 17:51:36 accel.accel_crc32c_C2 -- common/autotest_common.sh@10 -- # set +x 00:05:42.817 ************************************ 00:05:42.817 END TEST accel_crc32c_C2 00:05:42.817 ************************************ 00:05:42.817 17:51:36 accel -- common/autotest_common.sh@1142 -- # return 0 00:05:42.817 17:51:36 accel -- accel/accel.sh@103 -- # run_test accel_copy accel_test -t 1 -w copy -y 00:05:42.817 17:51:36 accel -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:05:42.817 17:51:36 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:42.817 17:51:36 accel -- common/autotest_common.sh@10 -- # set +x 00:05:42.817 ************************************ 00:05:42.817 START TEST accel_copy 00:05:42.817 ************************************ 00:05:42.817 17:51:36 accel.accel_copy -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w copy -y 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@16 -- # local accel_opc 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@17 -- # local accel_module 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@15 -- # accel_perf -t 1 -w copy -y 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w copy -y 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@12 -- # build_accel_config 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@40 -- # local IFS=, 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@41 -- # jq -r . 00:05:42.817 [2024-07-15 17:51:36.341019] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:42.817 [2024-07-15 17:51:36.341066] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid421001 ] 00:05:42.817 EAL: No free 2048 kB hugepages reported on node 1 00:05:42.817 [2024-07-15 17:51:36.395201] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:42.817 [2024-07-15 17:51:36.466348] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@20 -- # val=0x1 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@20 -- # val=copy 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@23 -- # accel_opc=copy 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@20 -- # val='4096 bytes' 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@20 -- # val=software 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@22 -- # accel_module=software 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@20 -- # val=32 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@20 -- # val=32 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@20 -- # val=1 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@20 -- # val='1 seconds' 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@20 -- # val=Yes 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:42.817 17:51:36 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@20 -- # val= 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@19 -- # IFS=: 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@19 -- # read -r var val 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@27 -- # [[ -n software ]] 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@27 -- # [[ -n copy ]] 00:05:44.193 17:51:37 accel.accel_copy -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:05:44.193 00:05:44.193 real 0m1.330s 00:05:44.193 user 0m1.237s 00:05:44.193 sys 0m0.106s 00:05:44.193 17:51:37 accel.accel_copy -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:44.193 17:51:37 accel.accel_copy -- common/autotest_common.sh@10 -- # set +x 00:05:44.193 ************************************ 00:05:44.193 END TEST accel_copy 00:05:44.193 ************************************ 00:05:44.193 17:51:37 accel -- common/autotest_common.sh@1142 -- # return 0 00:05:44.193 17:51:37 accel -- accel/accel.sh@104 -- # run_test accel_fill accel_test -t 1 -w fill -f 128 -q 64 -a 64 -y 00:05:44.193 17:51:37 accel -- common/autotest_common.sh@1099 -- # '[' 13 -le 1 ']' 00:05:44.193 17:51:37 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:44.193 17:51:37 accel -- common/autotest_common.sh@10 -- # set +x 00:05:44.193 ************************************ 00:05:44.193 START TEST accel_fill 00:05:44.193 ************************************ 00:05:44.193 17:51:37 accel.accel_fill -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w fill -f 128 -q 64 -a 64 -y 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@16 -- # local accel_opc 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@17 -- # local accel_module 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@15 -- # accel_perf -t 1 -w fill -f 128 -q 64 -a 64 -y 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w fill -f 128 -q 64 -a 64 -y 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@12 -- # build_accel_config 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@40 -- # local IFS=, 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@41 -- # jq -r . 00:05:44.193 [2024-07-15 17:51:37.730864] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:44.193 [2024-07-15 17:51:37.730912] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid421248 ] 00:05:44.193 EAL: No free 2048 kB hugepages reported on node 1 00:05:44.193 [2024-07-15 17:51:37.781695] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:44.193 [2024-07-15 17:51:37.853524] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:44.193 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@20 -- # val=0x1 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@20 -- # val=fill 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@23 -- # accel_opc=fill 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@20 -- # val=0x80 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@20 -- # val='4096 bytes' 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@20 -- # val=software 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@22 -- # accel_module=software 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@20 -- # val=64 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@20 -- # val=64 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@20 -- # val=1 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@20 -- # val='1 seconds' 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@20 -- # val=Yes 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:44.194 17:51:37 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@20 -- # val= 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@27 -- # [[ -n software ]] 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@27 -- # [[ -n fill ]] 00:05:45.581 17:51:39 accel.accel_fill -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:05:45.581 00:05:45.581 real 0m1.329s 00:05:45.581 user 0m1.234s 00:05:45.581 sys 0m0.107s 00:05:45.581 17:51:39 accel.accel_fill -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:45.581 17:51:39 accel.accel_fill -- common/autotest_common.sh@10 -- # set +x 00:05:45.581 ************************************ 00:05:45.581 END TEST accel_fill 00:05:45.581 ************************************ 00:05:45.581 17:51:39 accel -- common/autotest_common.sh@1142 -- # return 0 00:05:45.581 17:51:39 accel -- accel/accel.sh@105 -- # run_test accel_copy_crc32c accel_test -t 1 -w copy_crc32c -y 00:05:45.581 17:51:39 accel -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:05:45.581 17:51:39 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:45.581 17:51:39 accel -- common/autotest_common.sh@10 -- # set +x 00:05:45.581 ************************************ 00:05:45.581 START TEST accel_copy_crc32c 00:05:45.581 ************************************ 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w copy_crc32c -y 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@16 -- # local accel_opc 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@17 -- # local accel_module 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@15 -- # accel_perf -t 1 -w copy_crc32c -y 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w copy_crc32c -y 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@12 -- # build_accel_config 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@40 -- # local IFS=, 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@41 -- # jq -r . 00:05:45.581 [2024-07-15 17:51:39.108618] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:45.581 [2024-07-15 17:51:39.108663] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid421499 ] 00:05:45.581 EAL: No free 2048 kB hugepages reported on node 1 00:05:45.581 [2024-07-15 17:51:39.162333] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:45.581 [2024-07-15 17:51:39.238881] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val=0x1 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val=copy_crc32c 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@23 -- # accel_opc=copy_crc32c 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val=0 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val='4096 bytes' 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val='4096 bytes' 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val=software 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@22 -- # accel_module=software 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val=32 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val=32 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val=1 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val='1 seconds' 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val=Yes 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:45.581 17:51:39 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@20 -- # val= 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@27 -- # [[ -n software ]] 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@27 -- # [[ -n copy_crc32c ]] 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:05:46.960 00:05:46.960 real 0m1.337s 00:05:46.960 user 0m1.242s 00:05:46.960 sys 0m0.109s 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:46.960 17:51:40 accel.accel_copy_crc32c -- common/autotest_common.sh@10 -- # set +x 00:05:46.960 ************************************ 00:05:46.960 END TEST accel_copy_crc32c 00:05:46.960 ************************************ 00:05:46.960 17:51:40 accel -- common/autotest_common.sh@1142 -- # return 0 00:05:46.960 17:51:40 accel -- accel/accel.sh@106 -- # run_test accel_copy_crc32c_C2 accel_test -t 1 -w copy_crc32c -y -C 2 00:05:46.960 17:51:40 accel -- common/autotest_common.sh@1099 -- # '[' 9 -le 1 ']' 00:05:46.960 17:51:40 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:46.960 17:51:40 accel -- common/autotest_common.sh@10 -- # set +x 00:05:46.960 ************************************ 00:05:46.960 START TEST accel_copy_crc32c_C2 00:05:46.960 ************************************ 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w copy_crc32c -y -C 2 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@16 -- # local accel_opc 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@17 -- # local accel_module 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@15 -- # accel_perf -t 1 -w copy_crc32c -y -C 2 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w copy_crc32c -y -C 2 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@12 -- # build_accel_config 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@40 -- # local IFS=, 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@41 -- # jq -r . 00:05:46.960 [2024-07-15 17:51:40.500238] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:46.960 [2024-07-15 17:51:40.500292] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid421749 ] 00:05:46.960 EAL: No free 2048 kB hugepages reported on node 1 00:05:46.960 [2024-07-15 17:51:40.555749] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:46.960 [2024-07-15 17:51:40.627025] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val=0x1 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val=copy_crc32c 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@23 -- # accel_opc=copy_crc32c 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val=0 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val='4096 bytes' 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val='8192 bytes' 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val=software 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@22 -- # accel_module=software 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val=32 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val=32 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val=1 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val='1 seconds' 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val=Yes 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:46.960 17:51:40 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@20 -- # val= 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@21 -- # case "$var" in 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # IFS=: 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@19 -- # read -r var val 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@27 -- # [[ -n software ]] 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@27 -- # [[ -n copy_crc32c ]] 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:05:48.337 00:05:48.337 real 0m1.329s 00:05:48.337 user 0m1.234s 00:05:48.337 sys 0m0.110s 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:48.337 17:51:41 accel.accel_copy_crc32c_C2 -- common/autotest_common.sh@10 -- # set +x 00:05:48.337 ************************************ 00:05:48.337 END TEST accel_copy_crc32c_C2 00:05:48.337 ************************************ 00:05:48.337 17:51:41 accel -- common/autotest_common.sh@1142 -- # return 0 00:05:48.337 17:51:41 accel -- accel/accel.sh@107 -- # run_test accel_dualcast accel_test -t 1 -w dualcast -y 00:05:48.337 17:51:41 accel -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:05:48.337 17:51:41 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:48.337 17:51:41 accel -- common/autotest_common.sh@10 -- # set +x 00:05:48.337 ************************************ 00:05:48.337 START TEST accel_dualcast 00:05:48.337 ************************************ 00:05:48.337 17:51:41 accel.accel_dualcast -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w dualcast -y 00:05:48.337 17:51:41 accel.accel_dualcast -- accel/accel.sh@16 -- # local accel_opc 00:05:48.337 17:51:41 accel.accel_dualcast -- accel/accel.sh@17 -- # local accel_module 00:05:48.337 17:51:41 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:48.337 17:51:41 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:48.337 17:51:41 accel.accel_dualcast -- accel/accel.sh@15 -- # accel_perf -t 1 -w dualcast -y 00:05:48.337 17:51:41 accel.accel_dualcast -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w dualcast -y 00:05:48.337 17:51:41 accel.accel_dualcast -- accel/accel.sh@12 -- # build_accel_config 00:05:48.337 17:51:41 accel.accel_dualcast -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:48.337 17:51:41 accel.accel_dualcast -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:48.337 17:51:41 accel.accel_dualcast -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:48.337 17:51:41 accel.accel_dualcast -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:48.337 17:51:41 accel.accel_dualcast -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:48.337 17:51:41 accel.accel_dualcast -- accel/accel.sh@40 -- # local IFS=, 00:05:48.337 17:51:41 accel.accel_dualcast -- accel/accel.sh@41 -- # jq -r . 00:05:48.337 [2024-07-15 17:51:41.896632] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:48.337 [2024-07-15 17:51:41.896700] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid421994 ] 00:05:48.337 EAL: No free 2048 kB hugepages reported on node 1 00:05:48.337 [2024-07-15 17:51:41.953045] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:48.337 [2024-07-15 17:51:42.025389] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@20 -- # val=0x1 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@20 -- # val=dualcast 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@23 -- # accel_opc=dualcast 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@20 -- # val='4096 bytes' 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@20 -- # val=software 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@22 -- # accel_module=software 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@20 -- # val=32 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@20 -- # val=32 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:48.596 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@20 -- # val=1 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@20 -- # val='1 seconds' 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@20 -- # val=Yes 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:48.597 17:51:42 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@20 -- # val= 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@21 -- # case "$var" in 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@19 -- # IFS=: 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@19 -- # read -r var val 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@27 -- # [[ -n software ]] 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@27 -- # [[ -n dualcast ]] 00:05:49.534 17:51:43 accel.accel_dualcast -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:05:49.534 00:05:49.534 real 0m1.336s 00:05:49.534 user 0m1.234s 00:05:49.534 sys 0m0.114s 00:05:49.534 17:51:43 accel.accel_dualcast -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:49.534 17:51:43 accel.accel_dualcast -- common/autotest_common.sh@10 -- # set +x 00:05:49.534 ************************************ 00:05:49.534 END TEST accel_dualcast 00:05:49.534 ************************************ 00:05:49.534 17:51:43 accel -- common/autotest_common.sh@1142 -- # return 0 00:05:49.534 17:51:43 accel -- accel/accel.sh@108 -- # run_test accel_compare accel_test -t 1 -w compare -y 00:05:49.534 17:51:43 accel -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:05:49.534 17:51:43 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:49.534 17:51:43 accel -- common/autotest_common.sh@10 -- # set +x 00:05:49.793 ************************************ 00:05:49.793 START TEST accel_compare 00:05:49.793 ************************************ 00:05:49.793 17:51:43 accel.accel_compare -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w compare -y 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@16 -- # local accel_opc 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@17 -- # local accel_module 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@15 -- # accel_perf -t 1 -w compare -y 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w compare -y 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@12 -- # build_accel_config 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@40 -- # local IFS=, 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@41 -- # jq -r . 00:05:49.793 [2024-07-15 17:51:43.279594] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:49.793 [2024-07-15 17:51:43.279643] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid422251 ] 00:05:49.793 EAL: No free 2048 kB hugepages reported on node 1 00:05:49.793 [2024-07-15 17:51:43.333306] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:49.793 [2024-07-15 17:51:43.404230] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@20 -- # val=0x1 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@20 -- # val=compare 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@23 -- # accel_opc=compare 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@20 -- # val='4096 bytes' 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:49.793 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@20 -- # val=software 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@22 -- # accel_module=software 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@20 -- # val=32 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@20 -- # val=32 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@20 -- # val=1 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@20 -- # val='1 seconds' 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@20 -- # val=Yes 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:49.794 17:51:43 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@20 -- # val= 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@27 -- # [[ -n software ]] 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@27 -- # [[ -n compare ]] 00:05:51.173 17:51:44 accel.accel_compare -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:05:51.173 00:05:51.173 real 0m1.319s 00:05:51.173 user 0m1.226s 00:05:51.173 sys 0m0.106s 00:05:51.173 17:51:44 accel.accel_compare -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:51.173 17:51:44 accel.accel_compare -- common/autotest_common.sh@10 -- # set +x 00:05:51.173 ************************************ 00:05:51.173 END TEST accel_compare 00:05:51.173 ************************************ 00:05:51.173 17:51:44 accel -- common/autotest_common.sh@1142 -- # return 0 00:05:51.173 17:51:44 accel -- accel/accel.sh@109 -- # run_test accel_xor accel_test -t 1 -w xor -y 00:05:51.173 17:51:44 accel -- common/autotest_common.sh@1099 -- # '[' 7 -le 1 ']' 00:05:51.173 17:51:44 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:51.173 17:51:44 accel -- common/autotest_common.sh@10 -- # set +x 00:05:51.173 ************************************ 00:05:51.173 START TEST accel_xor 00:05:51.173 ************************************ 00:05:51.173 17:51:44 accel.accel_xor -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w xor -y 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@16 -- # local accel_opc 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@17 -- # local accel_module 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@15 -- # accel_perf -t 1 -w xor -y 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w xor -y 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@12 -- # build_accel_config 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@40 -- # local IFS=, 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@41 -- # jq -r . 00:05:51.173 [2024-07-15 17:51:44.662853] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:51.173 [2024-07-15 17:51:44.662915] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid422499 ] 00:05:51.173 EAL: No free 2048 kB hugepages reported on node 1 00:05:51.173 [2024-07-15 17:51:44.718676] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:51.173 [2024-07-15 17:51:44.789914] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@20 -- # val=0x1 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@20 -- # val=xor 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@23 -- # accel_opc=xor 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@20 -- # val=2 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@20 -- # val='4096 bytes' 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@20 -- # val=software 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@22 -- # accel_module=software 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@20 -- # val=32 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@20 -- # val=32 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@20 -- # val=1 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@20 -- # val='1 seconds' 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@20 -- # val=Yes 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:51.173 17:51:44 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@27 -- # [[ -n software ]] 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@27 -- # [[ -n xor ]] 00:05:52.553 17:51:45 accel.accel_xor -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:05:52.553 00:05:52.553 real 0m1.332s 00:05:52.553 user 0m1.239s 00:05:52.553 sys 0m0.106s 00:05:52.553 17:51:45 accel.accel_xor -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:52.553 17:51:45 accel.accel_xor -- common/autotest_common.sh@10 -- # set +x 00:05:52.553 ************************************ 00:05:52.553 END TEST accel_xor 00:05:52.553 ************************************ 00:05:52.553 17:51:45 accel -- common/autotest_common.sh@1142 -- # return 0 00:05:52.553 17:51:45 accel -- accel/accel.sh@110 -- # run_test accel_xor accel_test -t 1 -w xor -y -x 3 00:05:52.553 17:51:45 accel -- common/autotest_common.sh@1099 -- # '[' 9 -le 1 ']' 00:05:52.553 17:51:45 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:52.553 17:51:45 accel -- common/autotest_common.sh@10 -- # set +x 00:05:52.553 ************************************ 00:05:52.553 START TEST accel_xor 00:05:52.553 ************************************ 00:05:52.553 17:51:46 accel.accel_xor -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w xor -y -x 3 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@16 -- # local accel_opc 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@17 -- # local accel_module 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@15 -- # accel_perf -t 1 -w xor -y -x 3 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w xor -y -x 3 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@12 -- # build_accel_config 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@40 -- # local IFS=, 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@41 -- # jq -r . 00:05:52.553 [2024-07-15 17:51:46.041849] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:52.553 [2024-07-15 17:51:46.041897] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid422753 ] 00:05:52.553 EAL: No free 2048 kB hugepages reported on node 1 00:05:52.553 [2024-07-15 17:51:46.096089] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:52.553 [2024-07-15 17:51:46.167551] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@20 -- # val=0x1 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@20 -- # val=xor 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@23 -- # accel_opc=xor 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@20 -- # val=3 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@20 -- # val='4096 bytes' 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@20 -- # val=software 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@22 -- # accel_module=software 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@20 -- # val=32 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@20 -- # val=32 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@20 -- # val=1 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.553 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.554 17:51:46 accel.accel_xor -- accel/accel.sh@20 -- # val='1 seconds' 00:05:52.554 17:51:46 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.554 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.554 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.554 17:51:46 accel.accel_xor -- accel/accel.sh@20 -- # val=Yes 00:05:52.554 17:51:46 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.554 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.554 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.554 17:51:46 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:52.554 17:51:46 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.554 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.554 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:52.554 17:51:46 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:52.554 17:51:46 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:52.554 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:52.554 17:51:46 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:53.933 17:51:47 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:53.933 17:51:47 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:53.933 17:51:47 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:53.933 17:51:47 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:53.933 17:51:47 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:53.933 17:51:47 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:53.933 17:51:47 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:53.934 17:51:47 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:53.934 17:51:47 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:53.934 17:51:47 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_xor -- accel/accel.sh@20 -- # val= 00:05:53.934 17:51:47 accel.accel_xor -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_xor -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_xor -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_xor -- accel/accel.sh@27 -- # [[ -n software ]] 00:05:53.934 17:51:47 accel.accel_xor -- accel/accel.sh@27 -- # [[ -n xor ]] 00:05:53.934 17:51:47 accel.accel_xor -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:05:53.934 00:05:53.934 real 0m1.331s 00:05:53.934 user 0m1.239s 00:05:53.934 sys 0m0.106s 00:05:53.934 17:51:47 accel.accel_xor -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:53.934 17:51:47 accel.accel_xor -- common/autotest_common.sh@10 -- # set +x 00:05:53.934 ************************************ 00:05:53.934 END TEST accel_xor 00:05:53.934 ************************************ 00:05:53.934 17:51:47 accel -- common/autotest_common.sh@1142 -- # return 0 00:05:53.934 17:51:47 accel -- accel/accel.sh@111 -- # run_test accel_dif_verify accel_test -t 1 -w dif_verify 00:05:53.934 17:51:47 accel -- common/autotest_common.sh@1099 -- # '[' 6 -le 1 ']' 00:05:53.934 17:51:47 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:53.934 17:51:47 accel -- common/autotest_common.sh@10 -- # set +x 00:05:53.934 ************************************ 00:05:53.934 START TEST accel_dif_verify 00:05:53.934 ************************************ 00:05:53.934 17:51:47 accel.accel_dif_verify -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w dif_verify 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@16 -- # local accel_opc 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@17 -- # local accel_module 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@15 -- # accel_perf -t 1 -w dif_verify 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w dif_verify 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@12 -- # build_accel_config 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@40 -- # local IFS=, 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@41 -- # jq -r . 00:05:53.934 [2024-07-15 17:51:47.415297] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:53.934 [2024-07-15 17:51:47.415361] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid422998 ] 00:05:53.934 EAL: No free 2048 kB hugepages reported on node 1 00:05:53.934 [2024-07-15 17:51:47.470590] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:53.934 [2024-07-15 17:51:47.541923] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@20 -- # val=0x1 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@20 -- # val=dif_verify 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@23 -- # accel_opc=dif_verify 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@20 -- # val='4096 bytes' 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@20 -- # val='4096 bytes' 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@20 -- # val='512 bytes' 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@20 -- # val='8 bytes' 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@20 -- # val=software 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@22 -- # accel_module=software 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@20 -- # val=32 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@20 -- # val=32 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@20 -- # val=1 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@20 -- # val='1 seconds' 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@20 -- # val=No 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:53.934 17:51:47 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@20 -- # val= 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@21 -- # case "$var" in 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@19 -- # IFS=: 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@19 -- # read -r var val 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@27 -- # [[ -n software ]] 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@27 -- # [[ -n dif_verify ]] 00:05:55.316 17:51:48 accel.accel_dif_verify -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:05:55.316 00:05:55.316 real 0m1.334s 00:05:55.316 user 0m1.240s 00:05:55.316 sys 0m0.109s 00:05:55.316 17:51:48 accel.accel_dif_verify -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:55.316 17:51:48 accel.accel_dif_verify -- common/autotest_common.sh@10 -- # set +x 00:05:55.316 ************************************ 00:05:55.316 END TEST accel_dif_verify 00:05:55.316 ************************************ 00:05:55.316 17:51:48 accel -- common/autotest_common.sh@1142 -- # return 0 00:05:55.316 17:51:48 accel -- accel/accel.sh@112 -- # run_test accel_dif_generate accel_test -t 1 -w dif_generate 00:05:55.316 17:51:48 accel -- common/autotest_common.sh@1099 -- # '[' 6 -le 1 ']' 00:05:55.316 17:51:48 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:55.316 17:51:48 accel -- common/autotest_common.sh@10 -- # set +x 00:05:55.316 ************************************ 00:05:55.316 START TEST accel_dif_generate 00:05:55.316 ************************************ 00:05:55.316 17:51:48 accel.accel_dif_generate -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w dif_generate 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@16 -- # local accel_opc 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@17 -- # local accel_module 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@15 -- # accel_perf -t 1 -w dif_generate 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w dif_generate 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@12 -- # build_accel_config 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@40 -- # local IFS=, 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@41 -- # jq -r . 00:05:55.316 [2024-07-15 17:51:48.811304] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:55.316 [2024-07-15 17:51:48.811376] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid423245 ] 00:05:55.316 EAL: No free 2048 kB hugepages reported on node 1 00:05:55.316 [2024-07-15 17:51:48.867503] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:55.316 [2024-07-15 17:51:48.939756] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@20 -- # val=0x1 00:05:55.316 17:51:48 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@20 -- # val=dif_generate 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@23 -- # accel_opc=dif_generate 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@20 -- # val='4096 bytes' 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@20 -- # val='4096 bytes' 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@20 -- # val='512 bytes' 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@20 -- # val='8 bytes' 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@20 -- # val=software 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@22 -- # accel_module=software 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@20 -- # val=32 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@20 -- # val=32 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@20 -- # val=1 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@20 -- # val='1 seconds' 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@20 -- # val=No 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:55.317 17:51:48 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@20 -- # val= 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@21 -- # case "$var" in 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@19 -- # IFS=: 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@19 -- # read -r var val 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@27 -- # [[ -n software ]] 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@27 -- # [[ -n dif_generate ]] 00:05:56.695 17:51:50 accel.accel_dif_generate -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:05:56.695 00:05:56.695 real 0m1.339s 00:05:56.695 user 0m1.248s 00:05:56.695 sys 0m0.105s 00:05:56.695 17:51:50 accel.accel_dif_generate -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:56.695 17:51:50 accel.accel_dif_generate -- common/autotest_common.sh@10 -- # set +x 00:05:56.695 ************************************ 00:05:56.695 END TEST accel_dif_generate 00:05:56.695 ************************************ 00:05:56.695 17:51:50 accel -- common/autotest_common.sh@1142 -- # return 0 00:05:56.695 17:51:50 accel -- accel/accel.sh@113 -- # run_test accel_dif_generate_copy accel_test -t 1 -w dif_generate_copy 00:05:56.695 17:51:50 accel -- common/autotest_common.sh@1099 -- # '[' 6 -le 1 ']' 00:05:56.695 17:51:50 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:56.695 17:51:50 accel -- common/autotest_common.sh@10 -- # set +x 00:05:56.695 ************************************ 00:05:56.695 START TEST accel_dif_generate_copy 00:05:56.695 ************************************ 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w dif_generate_copy 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@16 -- # local accel_opc 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@17 -- # local accel_module 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@15 -- # accel_perf -t 1 -w dif_generate_copy 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w dif_generate_copy 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@12 -- # build_accel_config 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@40 -- # local IFS=, 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@41 -- # jq -r . 00:05:56.695 [2024-07-15 17:51:50.210110] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:56.695 [2024-07-15 17:51:50.210166] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid423500 ] 00:05:56.695 EAL: No free 2048 kB hugepages reported on node 1 00:05:56.695 [2024-07-15 17:51:50.267628] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:56.695 [2024-07-15 17:51:50.339757] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val=0x1 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val=dif_generate_copy 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@23 -- # accel_opc=dif_generate_copy 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val='4096 bytes' 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:56.695 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val='4096 bytes' 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val=software 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@22 -- # accel_module=software 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val=32 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val=32 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val=1 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val='1 seconds' 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val=No 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:56.696 17:51:50 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@20 -- # val= 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@21 -- # case "$var" in 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # IFS=: 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@19 -- # read -r var val 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@27 -- # [[ -n software ]] 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@27 -- # [[ -n dif_generate_copy ]] 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:05:58.071 00:05:58.071 real 0m1.337s 00:05:58.071 user 0m1.236s 00:05:58.071 sys 0m0.115s 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:58.071 17:51:51 accel.accel_dif_generate_copy -- common/autotest_common.sh@10 -- # set +x 00:05:58.071 ************************************ 00:05:58.071 END TEST accel_dif_generate_copy 00:05:58.071 ************************************ 00:05:58.071 17:51:51 accel -- common/autotest_common.sh@1142 -- # return 0 00:05:58.071 17:51:51 accel -- accel/accel.sh@115 -- # [[ y == y ]] 00:05:58.071 17:51:51 accel -- accel/accel.sh@116 -- # run_test accel_comp accel_test -t 1 -w compress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:05:58.071 17:51:51 accel -- common/autotest_common.sh@1099 -- # '[' 8 -le 1 ']' 00:05:58.071 17:51:51 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:58.071 17:51:51 accel -- common/autotest_common.sh@10 -- # set +x 00:05:58.071 ************************************ 00:05:58.071 START TEST accel_comp 00:05:58.071 ************************************ 00:05:58.071 17:51:51 accel.accel_comp -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w compress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@16 -- # local accel_opc 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@17 -- # local accel_module 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@15 -- # accel_perf -t 1 -w compress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w compress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@12 -- # build_accel_config 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@40 -- # local IFS=, 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@41 -- # jq -r . 00:05:58.071 [2024-07-15 17:51:51.582667] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:58.071 [2024-07-15 17:51:51.582702] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid423747 ] 00:05:58.071 EAL: No free 2048 kB hugepages reported on node 1 00:05:58.071 [2024-07-15 17:51:51.635599] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:58.071 [2024-07-15 17:51:51.707036] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:58.071 17:51:51 accel.accel_comp -- accel/accel.sh@20 -- # val=0x1 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@20 -- # val=compress 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@23 -- # accel_opc=compress 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@20 -- # val='4096 bytes' 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@20 -- # val=software 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@22 -- # accel_module=software 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@20 -- # val=32 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@20 -- # val=32 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@20 -- # val=1 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@20 -- # val='1 seconds' 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@20 -- # val=No 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:58.072 17:51:51 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@20 -- # val= 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@27 -- # [[ -n software ]] 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@27 -- # [[ -n compress ]] 00:05:59.511 17:51:52 accel.accel_comp -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:05:59.511 00:05:59.511 real 0m1.320s 00:05:59.511 user 0m1.229s 00:05:59.511 sys 0m0.105s 00:05:59.511 17:51:52 accel.accel_comp -- common/autotest_common.sh@1124 -- # xtrace_disable 00:05:59.511 17:51:52 accel.accel_comp -- common/autotest_common.sh@10 -- # set +x 00:05:59.511 ************************************ 00:05:59.511 END TEST accel_comp 00:05:59.511 ************************************ 00:05:59.511 17:51:52 accel -- common/autotest_common.sh@1142 -- # return 0 00:05:59.511 17:51:52 accel -- accel/accel.sh@117 -- # run_test accel_decomp accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y 00:05:59.511 17:51:52 accel -- common/autotest_common.sh@1099 -- # '[' 9 -le 1 ']' 00:05:59.511 17:51:52 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:05:59.511 17:51:52 accel -- common/autotest_common.sh@10 -- # set +x 00:05:59.511 ************************************ 00:05:59.511 START TEST accel_decomp 00:05:59.511 ************************************ 00:05:59.511 17:51:52 accel.accel_decomp -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y 00:05:59.511 17:51:52 accel.accel_decomp -- accel/accel.sh@16 -- # local accel_opc 00:05:59.511 17:51:52 accel.accel_decomp -- accel/accel.sh@17 -- # local accel_module 00:05:59.511 17:51:52 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:52 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:52 accel.accel_decomp -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y 00:05:59.511 17:51:52 accel.accel_decomp -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y 00:05:59.511 17:51:52 accel.accel_decomp -- accel/accel.sh@12 -- # build_accel_config 00:05:59.511 17:51:52 accel.accel_decomp -- accel/accel.sh@31 -- # accel_json_cfg=() 00:05:59.511 17:51:52 accel.accel_decomp -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:05:59.511 17:51:52 accel.accel_decomp -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:05:59.511 17:51:52 accel.accel_decomp -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:05:59.511 17:51:52 accel.accel_decomp -- accel/accel.sh@36 -- # [[ -n '' ]] 00:05:59.511 17:51:52 accel.accel_decomp -- accel/accel.sh@40 -- # local IFS=, 00:05:59.511 17:51:52 accel.accel_decomp -- accel/accel.sh@41 -- # jq -r . 00:05:59.511 [2024-07-15 17:51:52.976975] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:05:59.511 [2024-07-15 17:51:52.977025] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid423994 ] 00:05:59.511 EAL: No free 2048 kB hugepages reported on node 1 00:05:59.511 [2024-07-15 17:51:53.031294] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:05:59.511 [2024-07-15 17:51:53.102594] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@20 -- # val=0x1 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@20 -- # val=decompress 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@23 -- # accel_opc=decompress 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@20 -- # val='4096 bytes' 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@20 -- # val=software 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@22 -- # accel_module=software 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@20 -- # val=32 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@20 -- # val=32 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@20 -- # val=1 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@20 -- # val='1 seconds' 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@20 -- # val=Yes 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:05:59.511 17:51:53 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:05:59.512 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:05:59.512 17:51:53 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@20 -- # val= 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@27 -- # [[ -n software ]] 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:06:00.890 17:51:54 accel.accel_decomp -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:06:00.890 00:06:00.890 real 0m1.335s 00:06:00.890 user 0m1.233s 00:06:00.890 sys 0m0.116s 00:06:00.890 17:51:54 accel.accel_decomp -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:00.890 17:51:54 accel.accel_decomp -- common/autotest_common.sh@10 -- # set +x 00:06:00.890 ************************************ 00:06:00.890 END TEST accel_decomp 00:06:00.890 ************************************ 00:06:00.890 17:51:54 accel -- common/autotest_common.sh@1142 -- # return 0 00:06:00.890 17:51:54 accel -- accel/accel.sh@118 -- # run_test accel_decomp_full accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 00:06:00.890 17:51:54 accel -- common/autotest_common.sh@1099 -- # '[' 11 -le 1 ']' 00:06:00.890 17:51:54 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:00.890 17:51:54 accel -- common/autotest_common.sh@10 -- # set +x 00:06:00.890 ************************************ 00:06:00.890 START TEST accel_decomp_full 00:06:00.890 ************************************ 00:06:00.890 17:51:54 accel.accel_decomp_full -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@16 -- # local accel_opc 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@17 -- # local accel_module 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@12 -- # build_accel_config 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@31 -- # accel_json_cfg=() 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@36 -- # [[ -n '' ]] 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@40 -- # local IFS=, 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@41 -- # jq -r . 00:06:00.890 [2024-07-15 17:51:54.372259] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:06:00.890 [2024-07-15 17:51:54.372303] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid424248 ] 00:06:00.890 EAL: No free 2048 kB hugepages reported on node 1 00:06:00.890 [2024-07-15 17:51:54.426403] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:00.890 [2024-07-15 17:51:54.499389] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@20 -- # val=0x1 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@20 -- # val=decompress 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@23 -- # accel_opc=decompress 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@20 -- # val='111250 bytes' 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@20 -- # val=software 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@22 -- # accel_module=software 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@20 -- # val=32 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@20 -- # val=32 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@20 -- # val=1 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@20 -- # val='1 seconds' 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@20 -- # val=Yes 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:00.890 17:51:54 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@20 -- # val= 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@21 -- # case "$var" in 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@19 -- # IFS=: 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@19 -- # read -r var val 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@27 -- # [[ -n software ]] 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:06:02.270 17:51:55 accel.accel_decomp_full -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:06:02.270 00:06:02.270 real 0m1.345s 00:06:02.270 user 0m1.245s 00:06:02.270 sys 0m0.113s 00:06:02.270 17:51:55 accel.accel_decomp_full -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:02.270 17:51:55 accel.accel_decomp_full -- common/autotest_common.sh@10 -- # set +x 00:06:02.270 ************************************ 00:06:02.270 END TEST accel_decomp_full 00:06:02.270 ************************************ 00:06:02.270 17:51:55 accel -- common/autotest_common.sh@1142 -- # return 0 00:06:02.270 17:51:55 accel -- accel/accel.sh@119 -- # run_test accel_decomp_mcore accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -m 0xf 00:06:02.270 17:51:55 accel -- common/autotest_common.sh@1099 -- # '[' 11 -le 1 ']' 00:06:02.270 17:51:55 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:02.270 17:51:55 accel -- common/autotest_common.sh@10 -- # set +x 00:06:02.271 ************************************ 00:06:02.271 START TEST accel_decomp_mcore 00:06:02.271 ************************************ 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -m 0xf 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@16 -- # local accel_opc 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@17 -- # local accel_module 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -m 0xf 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -m 0xf 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@12 -- # build_accel_config 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@31 -- # accel_json_cfg=() 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@36 -- # [[ -n '' ]] 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@40 -- # local IFS=, 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@41 -- # jq -r . 00:06:02.271 [2024-07-15 17:51:55.779110] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:06:02.271 [2024-07-15 17:51:55.779154] [ DPDK EAL parameters: accel_perf --no-shconf -c 0xf --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid424499 ] 00:06:02.271 EAL: No free 2048 kB hugepages reported on node 1 00:06:02.271 [2024-07-15 17:51:55.833192] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:06:02.271 [2024-07-15 17:51:55.906929] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:06:02.271 [2024-07-15 17:51:55.906946] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:06:02.271 [2024-07-15 17:51:55.907031] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:06:02.271 [2024-07-15 17:51:55.907033] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val=0xf 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val=decompress 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@23 -- # accel_opc=decompress 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val='4096 bytes' 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val=software 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@22 -- # accel_module=software 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val=32 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val=32 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val=1 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val='1 seconds' 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val=Yes 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:02.271 17:51:55 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@20 -- # val= 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@27 -- # [[ -n software ]] 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:06:03.651 00:06:03.651 real 0m1.343s 00:06:03.651 user 0m4.559s 00:06:03.651 sys 0m0.124s 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:03.651 17:51:57 accel.accel_decomp_mcore -- common/autotest_common.sh@10 -- # set +x 00:06:03.651 ************************************ 00:06:03.651 END TEST accel_decomp_mcore 00:06:03.651 ************************************ 00:06:03.651 17:51:57 accel -- common/autotest_common.sh@1142 -- # return 0 00:06:03.651 17:51:57 accel -- accel/accel.sh@120 -- # run_test accel_decomp_full_mcore accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 -m 0xf 00:06:03.651 17:51:57 accel -- common/autotest_common.sh@1099 -- # '[' 13 -le 1 ']' 00:06:03.651 17:51:57 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:03.651 17:51:57 accel -- common/autotest_common.sh@10 -- # set +x 00:06:03.651 ************************************ 00:06:03.651 START TEST accel_decomp_full_mcore 00:06:03.651 ************************************ 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 -m 0xf 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@16 -- # local accel_opc 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@17 -- # local accel_module 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 -m 0xf 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 -m 0xf 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@12 -- # build_accel_config 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@31 -- # accel_json_cfg=() 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@36 -- # [[ -n '' ]] 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@40 -- # local IFS=, 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@41 -- # jq -r . 00:06:03.651 [2024-07-15 17:51:57.173579] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:06:03.651 [2024-07-15 17:51:57.173624] [ DPDK EAL parameters: accel_perf --no-shconf -c 0xf --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid424747 ] 00:06:03.651 EAL: No free 2048 kB hugepages reported on node 1 00:06:03.651 [2024-07-15 17:51:57.228013] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:06:03.651 [2024-07-15 17:51:57.302273] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:06:03.651 [2024-07-15 17:51:57.302384] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:06:03.651 [2024-07-15 17:51:57.302479] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:06:03.651 [2024-07-15 17:51:57.302480] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val=0xf 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val=decompress 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@23 -- # accel_opc=decompress 00:06:03.651 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val='111250 bytes' 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val=software 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@22 -- # accel_module=software 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val=32 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val=32 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val=1 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val='1 seconds' 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val=Yes 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:03.652 17:51:57 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@20 -- # val= 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@21 -- # case "$var" in 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # IFS=: 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@19 -- # read -r var val 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@27 -- # [[ -n software ]] 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:06:05.032 00:06:05.032 real 0m1.353s 00:06:05.032 user 0m4.598s 00:06:05.032 sys 0m0.118s 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:05.032 17:51:58 accel.accel_decomp_full_mcore -- common/autotest_common.sh@10 -- # set +x 00:06:05.032 ************************************ 00:06:05.032 END TEST accel_decomp_full_mcore 00:06:05.032 ************************************ 00:06:05.032 17:51:58 accel -- common/autotest_common.sh@1142 -- # return 0 00:06:05.032 17:51:58 accel -- accel/accel.sh@121 -- # run_test accel_decomp_mthread accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -T 2 00:06:05.032 17:51:58 accel -- common/autotest_common.sh@1099 -- # '[' 11 -le 1 ']' 00:06:05.032 17:51:58 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:05.032 17:51:58 accel -- common/autotest_common.sh@10 -- # set +x 00:06:05.032 ************************************ 00:06:05.032 START TEST accel_decomp_mthread 00:06:05.032 ************************************ 00:06:05.032 17:51:58 accel.accel_decomp_mthread -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -T 2 00:06:05.032 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@16 -- # local accel_opc 00:06:05.032 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@17 -- # local accel_module 00:06:05.032 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -T 2 00:06:05.032 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:05.032 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:05.032 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -T 2 00:06:05.032 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@12 -- # build_accel_config 00:06:05.032 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@31 -- # accel_json_cfg=() 00:06:05.032 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:06:05.032 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:06:05.032 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:06:05.032 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@36 -- # [[ -n '' ]] 00:06:05.032 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@40 -- # local IFS=, 00:06:05.032 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@41 -- # jq -r . 00:06:05.033 [2024-07-15 17:51:58.577543] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:06:05.033 [2024-07-15 17:51:58.577591] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid425006 ] 00:06:05.033 EAL: No free 2048 kB hugepages reported on node 1 00:06:05.033 [2024-07-15 17:51:58.631690] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:05.033 [2024-07-15 17:51:58.703487] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val=0x1 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val=decompress 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@23 -- # accel_opc=decompress 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val='4096 bytes' 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val=software 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@22 -- # accel_module=software 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val=32 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:05.033 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val=32 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val=2 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val='1 seconds' 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val=Yes 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:05.294 17:51:58 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@20 -- # val= 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@27 -- # [[ -n software ]] 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:06:06.232 00:06:06.232 real 0m1.328s 00:06:06.232 user 0m1.232s 00:06:06.232 sys 0m0.111s 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:06.232 17:51:59 accel.accel_decomp_mthread -- common/autotest_common.sh@10 -- # set +x 00:06:06.232 ************************************ 00:06:06.232 END TEST accel_decomp_mthread 00:06:06.232 ************************************ 00:06:06.232 17:51:59 accel -- common/autotest_common.sh@1142 -- # return 0 00:06:06.232 17:51:59 accel -- accel/accel.sh@122 -- # run_test accel_decomp_full_mthread accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 -T 2 00:06:06.232 17:51:59 accel -- common/autotest_common.sh@1099 -- # '[' 13 -le 1 ']' 00:06:06.232 17:51:59 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:06.232 17:51:59 accel -- common/autotest_common.sh@10 -- # set +x 00:06:06.232 ************************************ 00:06:06.232 START TEST accel_decomp_full_mthread 00:06:06.232 ************************************ 00:06:06.232 17:51:59 accel.accel_decomp_full_mthread -- common/autotest_common.sh@1123 -- # accel_test -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 -T 2 00:06:06.232 17:51:59 accel.accel_decomp_full_mthread -- accel/accel.sh@16 -- # local accel_opc 00:06:06.232 17:51:59 accel.accel_decomp_full_mthread -- accel/accel.sh@17 -- # local accel_module 00:06:06.232 17:51:59 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.232 17:51:59 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.232 17:51:59 accel.accel_decomp_full_mthread -- accel/accel.sh@15 -- # accel_perf -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 -T 2 00:06:06.232 17:51:59 accel.accel_decomp_full_mthread -- accel/accel.sh@12 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/accel_perf -c /dev/fd/62 -t 1 -w decompress -l /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib -y -o 0 -T 2 00:06:06.232 17:51:59 accel.accel_decomp_full_mthread -- accel/accel.sh@12 -- # build_accel_config 00:06:06.232 17:51:59 accel.accel_decomp_full_mthread -- accel/accel.sh@31 -- # accel_json_cfg=() 00:06:06.232 17:51:59 accel.accel_decomp_full_mthread -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:06:06.492 17:51:59 accel.accel_decomp_full_mthread -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:06:06.492 17:51:59 accel.accel_decomp_full_mthread -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:06:06.492 17:51:59 accel.accel_decomp_full_mthread -- accel/accel.sh@36 -- # [[ -n '' ]] 00:06:06.492 17:51:59 accel.accel_decomp_full_mthread -- accel/accel.sh@40 -- # local IFS=, 00:06:06.492 17:51:59 accel.accel_decomp_full_mthread -- accel/accel.sh@41 -- # jq -r . 00:06:06.492 [2024-07-15 17:51:59.979694] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:06:06.492 [2024-07-15 17:51:59.979746] [ DPDK EAL parameters: accel_perf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid425252 ] 00:06:06.492 EAL: No free 2048 kB hugepages reported on node 1 00:06:06.492 [2024-07-15 17:52:00.037328] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:06.492 [2024-07-15 17:52:00.116551] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val=0x1 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val=decompress 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@23 -- # accel_opc=decompress 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val='111250 bytes' 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val=software 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.492 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@22 -- # accel_module=software 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/bib 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val=32 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val=32 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val=2 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val='1 seconds' 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val=Yes 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:06.493 17:52:00 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@20 -- # val= 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@21 -- # case "$var" in 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # IFS=: 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@19 -- # read -r var val 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@27 -- # [[ -n software ]] 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@27 -- # [[ -n decompress ]] 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- accel/accel.sh@27 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:06:07.869 00:06:07.869 real 0m1.369s 00:06:07.869 user 0m1.266s 00:06:07.869 sys 0m0.118s 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:07.869 17:52:01 accel.accel_decomp_full_mthread -- common/autotest_common.sh@10 -- # set +x 00:06:07.869 ************************************ 00:06:07.869 END TEST accel_decomp_full_mthread 00:06:07.869 ************************************ 00:06:07.869 17:52:01 accel -- common/autotest_common.sh@1142 -- # return 0 00:06:07.869 17:52:01 accel -- accel/accel.sh@124 -- # [[ n == y ]] 00:06:07.869 17:52:01 accel -- accel/accel.sh@137 -- # run_test accel_dif_functional_tests /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/dif/dif -c /dev/fd/62 00:06:07.869 17:52:01 accel -- accel/accel.sh@137 -- # build_accel_config 00:06:07.869 17:52:01 accel -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:06:07.869 17:52:01 accel -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:07.869 17:52:01 accel -- accel/accel.sh@31 -- # accel_json_cfg=() 00:06:07.869 17:52:01 accel -- common/autotest_common.sh@10 -- # set +x 00:06:07.869 17:52:01 accel -- accel/accel.sh@32 -- # [[ 0 -gt 0 ]] 00:06:07.869 17:52:01 accel -- accel/accel.sh@33 -- # [[ 0 -gt 0 ]] 00:06:07.869 17:52:01 accel -- accel/accel.sh@34 -- # [[ 0 -gt 0 ]] 00:06:07.869 17:52:01 accel -- accel/accel.sh@36 -- # [[ -n '' ]] 00:06:07.869 17:52:01 accel -- accel/accel.sh@40 -- # local IFS=, 00:06:07.869 17:52:01 accel -- accel/accel.sh@41 -- # jq -r . 00:06:07.869 ************************************ 00:06:07.869 START TEST accel_dif_functional_tests 00:06:07.869 ************************************ 00:06:07.869 17:52:01 accel.accel_dif_functional_tests -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/dif/dif -c /dev/fd/62 00:06:07.869 [2024-07-15 17:52:01.414371] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:06:07.869 [2024-07-15 17:52:01.414409] [ DPDK EAL parameters: DIF --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid425506 ] 00:06:07.869 EAL: No free 2048 kB hugepages reported on node 1 00:06:07.869 [2024-07-15 17:52:01.467010] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:06:07.869 [2024-07-15 17:52:01.540285] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:06:07.869 [2024-07-15 17:52:01.540384] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:06:07.869 [2024-07-15 17:52:01.540386] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:06:08.128 00:06:08.128 00:06:08.128 CUnit - A unit testing framework for C - Version 2.1-3 00:06:08.128 http://cunit.sourceforge.net/ 00:06:08.128 00:06:08.128 00:06:08.128 Suite: accel_dif 00:06:08.128 Test: verify: DIF generated, GUARD check ...passed 00:06:08.128 Test: verify: DIF generated, APPTAG check ...passed 00:06:08.128 Test: verify: DIF generated, REFTAG check ...passed 00:06:08.128 Test: verify: DIF not generated, GUARD check ...[2024-07-15 17:52:01.608217] dif.c: 826:_dif_verify: *ERROR*: Failed to compare Guard: LBA=10, Expected=5a5a, Actual=7867 00:06:08.128 passed 00:06:08.128 Test: verify: DIF not generated, APPTAG check ...[2024-07-15 17:52:01.608269] dif.c: 841:_dif_verify: *ERROR*: Failed to compare App Tag: LBA=10, Expected=14, Actual=5a5a 00:06:08.128 passed 00:06:08.128 Test: verify: DIF not generated, REFTAG check ...[2024-07-15 17:52:01.608305] dif.c: 776:_dif_reftag_check: *ERROR*: Failed to compare Ref Tag: LBA=10, Expected=a, Actual=5a5a5a5a 00:06:08.128 passed 00:06:08.128 Test: verify: APPTAG correct, APPTAG check ...passed 00:06:08.128 Test: verify: APPTAG incorrect, APPTAG check ...[2024-07-15 17:52:01.608348] dif.c: 841:_dif_verify: *ERROR*: Failed to compare App Tag: LBA=30, Expected=28, Actual=14 00:06:08.128 passed 00:06:08.128 Test: verify: APPTAG incorrect, no APPTAG check ...passed 00:06:08.128 Test: verify: REFTAG incorrect, REFTAG ignore ...passed 00:06:08.128 Test: verify: REFTAG_INIT correct, REFTAG check ...passed 00:06:08.128 Test: verify: REFTAG_INIT incorrect, REFTAG check ...[2024-07-15 17:52:01.608446] dif.c: 776:_dif_reftag_check: *ERROR*: Failed to compare Ref Tag: LBA=10, Expected=a, Actual=10 00:06:08.128 passed 00:06:08.128 Test: verify copy: DIF generated, GUARD check ...passed 00:06:08.128 Test: verify copy: DIF generated, APPTAG check ...passed 00:06:08.128 Test: verify copy: DIF generated, REFTAG check ...passed 00:06:08.128 Test: verify copy: DIF not generated, GUARD check ...[2024-07-15 17:52:01.608550] dif.c: 826:_dif_verify: *ERROR*: Failed to compare Guard: LBA=10, Expected=5a5a, Actual=7867 00:06:08.128 passed 00:06:08.128 Test: verify copy: DIF not generated, APPTAG check ...[2024-07-15 17:52:01.608572] dif.c: 841:_dif_verify: *ERROR*: Failed to compare App Tag: LBA=10, Expected=14, Actual=5a5a 00:06:08.128 passed 00:06:08.128 Test: verify copy: DIF not generated, REFTAG check ...[2024-07-15 17:52:01.608590] dif.c: 776:_dif_reftag_check: *ERROR*: Failed to compare Ref Tag: LBA=10, Expected=a, Actual=5a5a5a5a 00:06:08.128 passed 00:06:08.128 Test: generate copy: DIF generated, GUARD check ...passed 00:06:08.128 Test: generate copy: DIF generated, APTTAG check ...passed 00:06:08.128 Test: generate copy: DIF generated, REFTAG check ...passed 00:06:08.128 Test: generate copy: DIF generated, no GUARD check flag set ...passed 00:06:08.128 Test: generate copy: DIF generated, no APPTAG check flag set ...passed 00:06:08.128 Test: generate copy: DIF generated, no REFTAG check flag set ...passed 00:06:08.128 Test: generate copy: iovecs-len validate ...[2024-07-15 17:52:01.608744] dif.c:1190:spdk_dif_generate_copy: *ERROR*: Size of bounce_iovs arrays are not valid or misaligned with block_size. 00:06:08.128 passed 00:06:08.128 Test: generate copy: buffer alignment validate ...passed 00:06:08.128 00:06:08.128 Run Summary: Type Total Ran Passed Failed Inactive 00:06:08.128 suites 1 1 n/a 0 0 00:06:08.128 tests 26 26 26 0 0 00:06:08.128 asserts 115 115 115 0 n/a 00:06:08.128 00:06:08.128 Elapsed time = 0.002 seconds 00:06:08.128 00:06:08.128 real 0m0.396s 00:06:08.128 user 0m0.615s 00:06:08.128 sys 0m0.136s 00:06:08.128 17:52:01 accel.accel_dif_functional_tests -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:08.128 17:52:01 accel.accel_dif_functional_tests -- common/autotest_common.sh@10 -- # set +x 00:06:08.128 ************************************ 00:06:08.128 END TEST accel_dif_functional_tests 00:06:08.128 ************************************ 00:06:08.128 17:52:01 accel -- common/autotest_common.sh@1142 -- # return 0 00:06:08.128 00:06:08.128 real 0m30.682s 00:06:08.128 user 0m34.740s 00:06:08.128 sys 0m3.969s 00:06:08.128 17:52:01 accel -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:08.128 17:52:01 accel -- common/autotest_common.sh@10 -- # set +x 00:06:08.128 ************************************ 00:06:08.128 END TEST accel 00:06:08.128 ************************************ 00:06:08.128 17:52:01 -- common/autotest_common.sh@1142 -- # return 0 00:06:08.128 17:52:01 -- spdk/autotest.sh@184 -- # run_test accel_rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/accel_rpc.sh 00:06:08.128 17:52:01 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:08.128 17:52:01 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:08.128 17:52:01 -- common/autotest_common.sh@10 -- # set +x 00:06:08.389 ************************************ 00:06:08.389 START TEST accel_rpc 00:06:08.389 ************************************ 00:06:08.389 17:52:01 accel_rpc -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel/accel_rpc.sh 00:06:08.389 * Looking for test storage... 00:06:08.389 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/accel 00:06:08.389 17:52:01 accel_rpc -- accel/accel_rpc.sh@11 -- # trap 'killprocess $spdk_tgt_pid; exit 1' ERR 00:06:08.389 17:52:01 accel_rpc -- accel/accel_rpc.sh@14 -- # spdk_tgt_pid=425728 00:06:08.389 17:52:01 accel_rpc -- accel/accel_rpc.sh@15 -- # waitforlisten 425728 00:06:08.389 17:52:01 accel_rpc -- accel/accel_rpc.sh@13 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --wait-for-rpc 00:06:08.389 17:52:01 accel_rpc -- common/autotest_common.sh@829 -- # '[' -z 425728 ']' 00:06:08.389 17:52:01 accel_rpc -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:08.389 17:52:01 accel_rpc -- common/autotest_common.sh@834 -- # local max_retries=100 00:06:08.389 17:52:01 accel_rpc -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:08.389 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:08.389 17:52:01 accel_rpc -- common/autotest_common.sh@838 -- # xtrace_disable 00:06:08.389 17:52:01 accel_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:08.389 [2024-07-15 17:52:01.969212] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:06:08.389 [2024-07-15 17:52:01.969267] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid425728 ] 00:06:08.389 EAL: No free 2048 kB hugepages reported on node 1 00:06:08.389 [2024-07-15 17:52:02.022458] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:08.389 [2024-07-15 17:52:02.103246] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:06:09.323 17:52:02 accel_rpc -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:06:09.323 17:52:02 accel_rpc -- common/autotest_common.sh@862 -- # return 0 00:06:09.323 17:52:02 accel_rpc -- accel/accel_rpc.sh@45 -- # [[ y == y ]] 00:06:09.323 17:52:02 accel_rpc -- accel/accel_rpc.sh@45 -- # [[ 0 -gt 0 ]] 00:06:09.323 17:52:02 accel_rpc -- accel/accel_rpc.sh@49 -- # [[ y == y ]] 00:06:09.323 17:52:02 accel_rpc -- accel/accel_rpc.sh@49 -- # [[ 0 -gt 0 ]] 00:06:09.323 17:52:02 accel_rpc -- accel/accel_rpc.sh@53 -- # run_test accel_assign_opcode accel_assign_opcode_test_suite 00:06:09.323 17:52:02 accel_rpc -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:09.323 17:52:02 accel_rpc -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:09.323 17:52:02 accel_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:09.323 ************************************ 00:06:09.323 START TEST accel_assign_opcode 00:06:09.323 ************************************ 00:06:09.323 17:52:02 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@1123 -- # accel_assign_opcode_test_suite 00:06:09.323 17:52:02 accel_rpc.accel_assign_opcode -- accel/accel_rpc.sh@38 -- # rpc_cmd accel_assign_opc -o copy -m incorrect 00:06:09.323 17:52:02 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:09.323 17:52:02 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@10 -- # set +x 00:06:09.323 [2024-07-15 17:52:02.813359] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation copy will be assigned to module incorrect 00:06:09.323 17:52:02 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:09.323 17:52:02 accel_rpc.accel_assign_opcode -- accel/accel_rpc.sh@40 -- # rpc_cmd accel_assign_opc -o copy -m software 00:06:09.323 17:52:02 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:09.323 17:52:02 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@10 -- # set +x 00:06:09.323 [2024-07-15 17:52:02.821372] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation copy will be assigned to module software 00:06:09.323 17:52:02 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:09.324 17:52:02 accel_rpc.accel_assign_opcode -- accel/accel_rpc.sh@41 -- # rpc_cmd framework_start_init 00:06:09.324 17:52:02 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:09.324 17:52:02 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@10 -- # set +x 00:06:09.324 17:52:03 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:09.324 17:52:03 accel_rpc.accel_assign_opcode -- accel/accel_rpc.sh@42 -- # rpc_cmd accel_get_opc_assignments 00:06:09.324 17:52:03 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:09.324 17:52:03 accel_rpc.accel_assign_opcode -- accel/accel_rpc.sh@42 -- # jq -r .copy 00:06:09.324 17:52:03 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@10 -- # set +x 00:06:09.324 17:52:03 accel_rpc.accel_assign_opcode -- accel/accel_rpc.sh@42 -- # grep software 00:06:09.324 17:52:03 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:09.324 software 00:06:09.324 00:06:09.324 real 0m0.235s 00:06:09.324 user 0m0.040s 00:06:09.324 sys 0m0.011s 00:06:09.324 17:52:03 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:09.324 17:52:03 accel_rpc.accel_assign_opcode -- common/autotest_common.sh@10 -- # set +x 00:06:09.324 ************************************ 00:06:09.324 END TEST accel_assign_opcode 00:06:09.324 ************************************ 00:06:09.584 17:52:03 accel_rpc -- common/autotest_common.sh@1142 -- # return 0 00:06:09.584 17:52:03 accel_rpc -- accel/accel_rpc.sh@55 -- # killprocess 425728 00:06:09.584 17:52:03 accel_rpc -- common/autotest_common.sh@948 -- # '[' -z 425728 ']' 00:06:09.584 17:52:03 accel_rpc -- common/autotest_common.sh@952 -- # kill -0 425728 00:06:09.584 17:52:03 accel_rpc -- common/autotest_common.sh@953 -- # uname 00:06:09.584 17:52:03 accel_rpc -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:06:09.584 17:52:03 accel_rpc -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 425728 00:06:09.584 17:52:03 accel_rpc -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:06:09.584 17:52:03 accel_rpc -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:06:09.585 17:52:03 accel_rpc -- common/autotest_common.sh@966 -- # echo 'killing process with pid 425728' 00:06:09.585 killing process with pid 425728 00:06:09.585 17:52:03 accel_rpc -- common/autotest_common.sh@967 -- # kill 425728 00:06:09.585 17:52:03 accel_rpc -- common/autotest_common.sh@972 -- # wait 425728 00:06:09.844 00:06:09.844 real 0m1.558s 00:06:09.844 user 0m1.640s 00:06:09.844 sys 0m0.396s 00:06:09.844 17:52:03 accel_rpc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:09.844 17:52:03 accel_rpc -- common/autotest_common.sh@10 -- # set +x 00:06:09.844 ************************************ 00:06:09.844 END TEST accel_rpc 00:06:09.844 ************************************ 00:06:09.844 17:52:03 -- common/autotest_common.sh@1142 -- # return 0 00:06:09.844 17:52:03 -- spdk/autotest.sh@185 -- # run_test app_cmdline /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/cmdline.sh 00:06:09.844 17:52:03 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:09.844 17:52:03 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:09.844 17:52:03 -- common/autotest_common.sh@10 -- # set +x 00:06:09.844 ************************************ 00:06:09.844 START TEST app_cmdline 00:06:09.844 ************************************ 00:06:09.844 17:52:03 app_cmdline -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/cmdline.sh 00:06:09.844 * Looking for test storage... 00:06:09.844 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app 00:06:09.844 17:52:03 app_cmdline -- app/cmdline.sh@14 -- # trap 'killprocess $spdk_tgt_pid' EXIT 00:06:09.844 17:52:03 app_cmdline -- app/cmdline.sh@17 -- # spdk_tgt_pid=426094 00:06:09.844 17:52:03 app_cmdline -- app/cmdline.sh@18 -- # waitforlisten 426094 00:06:09.844 17:52:03 app_cmdline -- app/cmdline.sh@16 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt --rpcs-allowed spdk_get_version,rpc_get_methods 00:06:09.844 17:52:03 app_cmdline -- common/autotest_common.sh@829 -- # '[' -z 426094 ']' 00:06:09.844 17:52:03 app_cmdline -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:09.844 17:52:03 app_cmdline -- common/autotest_common.sh@834 -- # local max_retries=100 00:06:09.844 17:52:03 app_cmdline -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:09.844 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:09.844 17:52:03 app_cmdline -- common/autotest_common.sh@838 -- # xtrace_disable 00:06:09.844 17:52:03 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:06:10.102 [2024-07-15 17:52:03.591132] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:06:10.102 [2024-07-15 17:52:03.591178] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid426094 ] 00:06:10.102 EAL: No free 2048 kB hugepages reported on node 1 00:06:10.102 [2024-07-15 17:52:03.644433] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:06:10.102 [2024-07-15 17:52:03.718380] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:06:10.672 17:52:04 app_cmdline -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:06:10.672 17:52:04 app_cmdline -- common/autotest_common.sh@862 -- # return 0 00:06:10.672 17:52:04 app_cmdline -- app/cmdline.sh@20 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py spdk_get_version 00:06:10.931 { 00:06:10.931 "version": "SPDK v24.09-pre git sha1 2f3522da7", 00:06:10.931 "fields": { 00:06:10.931 "major": 24, 00:06:10.931 "minor": 9, 00:06:10.931 "patch": 0, 00:06:10.931 "suffix": "-pre", 00:06:10.931 "commit": "2f3522da7" 00:06:10.931 } 00:06:10.931 } 00:06:10.931 17:52:04 app_cmdline -- app/cmdline.sh@22 -- # expected_methods=() 00:06:10.931 17:52:04 app_cmdline -- app/cmdline.sh@23 -- # expected_methods+=("rpc_get_methods") 00:06:10.931 17:52:04 app_cmdline -- app/cmdline.sh@24 -- # expected_methods+=("spdk_get_version") 00:06:10.931 17:52:04 app_cmdline -- app/cmdline.sh@26 -- # methods=($(rpc_cmd rpc_get_methods | jq -r ".[]" | sort)) 00:06:10.931 17:52:04 app_cmdline -- app/cmdline.sh@26 -- # rpc_cmd rpc_get_methods 00:06:10.931 17:52:04 app_cmdline -- app/cmdline.sh@26 -- # jq -r '.[]' 00:06:10.931 17:52:04 app_cmdline -- app/cmdline.sh@26 -- # sort 00:06:10.931 17:52:04 app_cmdline -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:10.931 17:52:04 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:06:10.931 17:52:04 app_cmdline -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:10.931 17:52:04 app_cmdline -- app/cmdline.sh@27 -- # (( 2 == 2 )) 00:06:10.931 17:52:04 app_cmdline -- app/cmdline.sh@28 -- # [[ rpc_get_methods spdk_get_version == \r\p\c\_\g\e\t\_\m\e\t\h\o\d\s\ \s\p\d\k\_\g\e\t\_\v\e\r\s\i\o\n ]] 00:06:10.931 17:52:04 app_cmdline -- app/cmdline.sh@30 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:06:10.931 17:52:04 app_cmdline -- common/autotest_common.sh@648 -- # local es=0 00:06:10.931 17:52:04 app_cmdline -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:06:10.931 17:52:04 app_cmdline -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:06:10.931 17:52:04 app_cmdline -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:06:10.931 17:52:04 app_cmdline -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:06:10.931 17:52:04 app_cmdline -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:06:10.931 17:52:04 app_cmdline -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:06:10.931 17:52:04 app_cmdline -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:06:10.931 17:52:04 app_cmdline -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:06:10.931 17:52:04 app_cmdline -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:06:10.931 17:52:04 app_cmdline -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py env_dpdk_get_mem_stats 00:06:11.190 request: 00:06:11.190 { 00:06:11.190 "method": "env_dpdk_get_mem_stats", 00:06:11.190 "req_id": 1 00:06:11.190 } 00:06:11.190 Got JSON-RPC error response 00:06:11.190 response: 00:06:11.190 { 00:06:11.190 "code": -32601, 00:06:11.190 "message": "Method not found" 00:06:11.190 } 00:06:11.190 17:52:04 app_cmdline -- common/autotest_common.sh@651 -- # es=1 00:06:11.190 17:52:04 app_cmdline -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:06:11.190 17:52:04 app_cmdline -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:06:11.190 17:52:04 app_cmdline -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:06:11.190 17:52:04 app_cmdline -- app/cmdline.sh@1 -- # killprocess 426094 00:06:11.190 17:52:04 app_cmdline -- common/autotest_common.sh@948 -- # '[' -z 426094 ']' 00:06:11.190 17:52:04 app_cmdline -- common/autotest_common.sh@952 -- # kill -0 426094 00:06:11.190 17:52:04 app_cmdline -- common/autotest_common.sh@953 -- # uname 00:06:11.190 17:52:04 app_cmdline -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:06:11.190 17:52:04 app_cmdline -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 426094 00:06:11.190 17:52:04 app_cmdline -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:06:11.190 17:52:04 app_cmdline -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:06:11.190 17:52:04 app_cmdline -- common/autotest_common.sh@966 -- # echo 'killing process with pid 426094' 00:06:11.190 killing process with pid 426094 00:06:11.190 17:52:04 app_cmdline -- common/autotest_common.sh@967 -- # kill 426094 00:06:11.190 17:52:04 app_cmdline -- common/autotest_common.sh@972 -- # wait 426094 00:06:11.448 00:06:11.448 real 0m1.643s 00:06:11.448 user 0m1.983s 00:06:11.448 sys 0m0.402s 00:06:11.448 17:52:05 app_cmdline -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:11.448 17:52:05 app_cmdline -- common/autotest_common.sh@10 -- # set +x 00:06:11.448 ************************************ 00:06:11.448 END TEST app_cmdline 00:06:11.448 ************************************ 00:06:11.448 17:52:05 -- common/autotest_common.sh@1142 -- # return 0 00:06:11.448 17:52:05 -- spdk/autotest.sh@186 -- # run_test version /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/version.sh 00:06:11.448 17:52:05 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:06:11.448 17:52:05 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:11.448 17:52:05 -- common/autotest_common.sh@10 -- # set +x 00:06:11.707 ************************************ 00:06:11.707 START TEST version 00:06:11.707 ************************************ 00:06:11.707 17:52:05 version -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/version.sh 00:06:11.707 * Looking for test storage... 00:06:11.707 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app 00:06:11.707 17:52:05 version -- app/version.sh@17 -- # get_header_version major 00:06:11.707 17:52:05 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MAJOR[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:06:11.707 17:52:05 version -- app/version.sh@14 -- # cut -f2 00:06:11.707 17:52:05 version -- app/version.sh@14 -- # tr -d '"' 00:06:11.707 17:52:05 version -- app/version.sh@17 -- # major=24 00:06:11.707 17:52:05 version -- app/version.sh@18 -- # get_header_version minor 00:06:11.707 17:52:05 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_MINOR[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:06:11.707 17:52:05 version -- app/version.sh@14 -- # cut -f2 00:06:11.707 17:52:05 version -- app/version.sh@14 -- # tr -d '"' 00:06:11.707 17:52:05 version -- app/version.sh@18 -- # minor=9 00:06:11.707 17:52:05 version -- app/version.sh@19 -- # get_header_version patch 00:06:11.707 17:52:05 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_PATCH[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:06:11.707 17:52:05 version -- app/version.sh@14 -- # cut -f2 00:06:11.707 17:52:05 version -- app/version.sh@14 -- # tr -d '"' 00:06:11.707 17:52:05 version -- app/version.sh@19 -- # patch=0 00:06:11.707 17:52:05 version -- app/version.sh@20 -- # get_header_version suffix 00:06:11.707 17:52:05 version -- app/version.sh@13 -- # grep -E '^#define SPDK_VERSION_SUFFIX[[:space:]]+' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/version.h 00:06:11.707 17:52:05 version -- app/version.sh@14 -- # cut -f2 00:06:11.707 17:52:05 version -- app/version.sh@14 -- # tr -d '"' 00:06:11.707 17:52:05 version -- app/version.sh@20 -- # suffix=-pre 00:06:11.707 17:52:05 version -- app/version.sh@22 -- # version=24.9 00:06:11.707 17:52:05 version -- app/version.sh@25 -- # (( patch != 0 )) 00:06:11.707 17:52:05 version -- app/version.sh@28 -- # version=24.9rc0 00:06:11.707 17:52:05 version -- app/version.sh@30 -- # PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python 00:06:11.707 17:52:05 version -- app/version.sh@30 -- # python3 -c 'import spdk; print(spdk.__version__)' 00:06:11.707 17:52:05 version -- app/version.sh@30 -- # py_version=24.9rc0 00:06:11.707 17:52:05 version -- app/version.sh@31 -- # [[ 24.9rc0 == \2\4\.\9\r\c\0 ]] 00:06:11.707 00:06:11.707 real 0m0.150s 00:06:11.707 user 0m0.086s 00:06:11.707 sys 0m0.100s 00:06:11.707 17:52:05 version -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:11.707 17:52:05 version -- common/autotest_common.sh@10 -- # set +x 00:06:11.707 ************************************ 00:06:11.707 END TEST version 00:06:11.707 ************************************ 00:06:11.707 17:52:05 -- common/autotest_common.sh@1142 -- # return 0 00:06:11.707 17:52:05 -- spdk/autotest.sh@188 -- # '[' 0 -eq 1 ']' 00:06:11.708 17:52:05 -- spdk/autotest.sh@198 -- # uname -s 00:06:11.708 17:52:05 -- spdk/autotest.sh@198 -- # [[ Linux == Linux ]] 00:06:11.708 17:52:05 -- spdk/autotest.sh@199 -- # [[ 0 -eq 1 ]] 00:06:11.708 17:52:05 -- spdk/autotest.sh@199 -- # [[ 0 -eq 1 ]] 00:06:11.708 17:52:05 -- spdk/autotest.sh@211 -- # '[' 0 -eq 1 ']' 00:06:11.708 17:52:05 -- spdk/autotest.sh@256 -- # '[' 0 -eq 1 ']' 00:06:11.708 17:52:05 -- spdk/autotest.sh@260 -- # timing_exit lib 00:06:11.708 17:52:05 -- common/autotest_common.sh@728 -- # xtrace_disable 00:06:11.708 17:52:05 -- common/autotest_common.sh@10 -- # set +x 00:06:11.708 17:52:05 -- spdk/autotest.sh@262 -- # '[' 0 -eq 1 ']' 00:06:11.708 17:52:05 -- spdk/autotest.sh@270 -- # '[' 0 -eq 1 ']' 00:06:11.708 17:52:05 -- spdk/autotest.sh@279 -- # '[' 1 -eq 1 ']' 00:06:11.708 17:52:05 -- spdk/autotest.sh@280 -- # export NET_TYPE 00:06:11.708 17:52:05 -- spdk/autotest.sh@283 -- # '[' tcp = rdma ']' 00:06:11.708 17:52:05 -- spdk/autotest.sh@286 -- # '[' tcp = tcp ']' 00:06:11.708 17:52:05 -- spdk/autotest.sh@287 -- # run_test nvmf_tcp /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf.sh --transport=tcp 00:06:11.708 17:52:05 -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:06:11.708 17:52:05 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:11.708 17:52:05 -- common/autotest_common.sh@10 -- # set +x 00:06:11.966 ************************************ 00:06:11.966 START TEST nvmf_tcp 00:06:11.966 ************************************ 00:06:11.966 17:52:05 nvmf_tcp -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/nvmf.sh --transport=tcp 00:06:11.966 * Looking for test storage... 00:06:11.966 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/nvmf.sh@10 -- # uname -s 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/nvmf.sh@10 -- # '[' '!' Linux = Linux ']' 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/nvmf.sh@14 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@7 -- # uname -s 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:06:11.966 17:52:05 nvmf_tcp -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:11.966 17:52:05 nvmf_tcp -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:11.966 17:52:05 nvmf_tcp -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:11.966 17:52:05 nvmf_tcp -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:11.966 17:52:05 nvmf_tcp -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:11.966 17:52:05 nvmf_tcp -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:11.966 17:52:05 nvmf_tcp -- paths/export.sh@5 -- # export PATH 00:06:11.966 17:52:05 nvmf_tcp -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@47 -- # : 0 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/common.sh@51 -- # have_pci_nics=0 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/nvmf.sh@16 -- # trap 'exit 1' SIGINT SIGTERM EXIT 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/nvmf.sh@18 -- # TEST_ARGS=("$@") 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/nvmf.sh@20 -- # timing_enter target 00:06:11.966 17:52:05 nvmf_tcp -- common/autotest_common.sh@722 -- # xtrace_disable 00:06:11.966 17:52:05 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/nvmf.sh@22 -- # [[ 0 -eq 0 ]] 00:06:11.966 17:52:05 nvmf_tcp -- nvmf/nvmf.sh@23 -- # run_test nvmf_example /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_example.sh --transport=tcp 00:06:11.966 17:52:05 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:06:11.966 17:52:05 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:11.966 17:52:05 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:06:11.966 ************************************ 00:06:11.966 START TEST nvmf_example 00:06:11.966 ************************************ 00:06:11.966 17:52:05 nvmf_tcp.nvmf_example -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_example.sh --transport=tcp 00:06:11.966 * Looking for test storage... 00:06:11.966 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:06:11.966 17:52:05 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:06:11.966 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@7 -- # uname -s 00:06:11.966 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:11.966 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:11.966 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:11.966 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:11.966 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:11.966 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:11.966 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:11.966 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:11.966 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:11.966 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:12.224 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:06:12.224 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:06:12.224 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:12.224 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:12.224 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:06:12.224 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:12.224 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:06:12.224 17:52:05 nvmf_tcp.nvmf_example -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:12.224 17:52:05 nvmf_tcp.nvmf_example -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:12.224 17:52:05 nvmf_tcp.nvmf_example -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:12.224 17:52:05 nvmf_tcp.nvmf_example -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:12.224 17:52:05 nvmf_tcp.nvmf_example -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:12.224 17:52:05 nvmf_tcp.nvmf_example -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:12.224 17:52:05 nvmf_tcp.nvmf_example -- paths/export.sh@5 -- # export PATH 00:06:12.224 17:52:05 nvmf_tcp.nvmf_example -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@47 -- # : 0 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@51 -- # have_pci_nics=0 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@11 -- # NVMF_EXAMPLE=("$SPDK_EXAMPLE_DIR/nvmf") 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@13 -- # MALLOC_BDEV_SIZE=64 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@14 -- # MALLOC_BLOCK_SIZE=512 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@24 -- # build_nvmf_example_args 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@17 -- # '[' 0 -eq 1 ']' 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@20 -- # NVMF_EXAMPLE+=(-i "$NVMF_APP_SHM_ID" -g 10000) 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@21 -- # NVMF_EXAMPLE+=("${NO_HUGE[@]}") 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@40 -- # timing_enter nvmf_example_test 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- common/autotest_common.sh@722 -- # xtrace_disable 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@41 -- # nvmftestinit 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@448 -- # prepare_net_devs 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@410 -- # local -g is_hw=no 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@412 -- # remove_spdk_ns 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- nvmf/common.sh@285 -- # xtrace_disable 00:06:12.225 17:52:05 nvmf_tcp.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:06:17.508 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:06:17.508 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@291 -- # pci_devs=() 00:06:17.508 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@291 -- # local -a pci_devs 00:06:17.508 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@292 -- # pci_net_devs=() 00:06:17.508 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:06:17.508 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@293 -- # pci_drivers=() 00:06:17.508 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@293 -- # local -A pci_drivers 00:06:17.508 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@295 -- # net_devs=() 00:06:17.508 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@295 -- # local -ga net_devs 00:06:17.508 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@296 -- # e810=() 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@296 -- # local -ga e810 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@297 -- # x722=() 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@297 -- # local -ga x722 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@298 -- # mlx=() 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@298 -- # local -ga mlx 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:06:17.509 Found 0000:86:00.0 (0x8086 - 0x159b) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:06:17.509 Found 0000:86:00.1 (0x8086 - 0x159b) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@390 -- # [[ up == up ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:06:17.509 Found net devices under 0000:86:00.0: cvl_0_0 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@390 -- # [[ up == up ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:06:17.509 Found net devices under 0000:86:00.1: cvl_0_1 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@414 -- # is_hw=yes 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:06:17.509 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:06:17.509 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.195 ms 00:06:17.509 00:06:17.509 --- 10.0.0.2 ping statistics --- 00:06:17.509 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:06:17.509 rtt min/avg/max/mdev = 0.195/0.195/0.195/0.000 ms 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:06:17.509 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:06:17.509 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.257 ms 00:06:17.509 00:06:17.509 --- 10.0.0.1 ping statistics --- 00:06:17.509 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:06:17.509 rtt min/avg/max/mdev = 0.257/0.257/0.257/0.000 ms 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@422 -- # return 0 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@42 -- # nvmfexamplestart '-m 0xF' 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@27 -- # timing_enter start_nvmf_example 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- common/autotest_common.sh@722 -- # xtrace_disable 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@29 -- # '[' tcp == tcp ']' 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@30 -- # NVMF_EXAMPLE=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_EXAMPLE[@]}") 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@33 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/nvmf -i 0 -g 10000 -m 0xF 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@34 -- # nvmfpid=429488 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@35 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@36 -- # waitforlisten 429488 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- common/autotest_common.sh@829 -- # '[' -z 429488 ']' 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- common/autotest_common.sh@834 -- # local max_retries=100 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:17.509 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- common/autotest_common.sh@838 -- # xtrace_disable 00:06:17.509 17:52:10 nvmf_tcp.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:06:17.509 EAL: No free 2048 kB hugepages reported on node 1 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- common/autotest_common.sh@862 -- # return 0 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@37 -- # timing_exit start_nvmf_example 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- common/autotest_common.sh@728 -- # xtrace_disable 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@45 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@47 -- # rpc_cmd bdev_malloc_create 64 512 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@47 -- # malloc_bdevs='Malloc0 ' 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@49 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@52 -- # for malloc_bdev in $malloc_bdevs 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@53 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@57 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@59 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:06:18.077 17:52:11 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 64 -o 4096 -w randrw -M 30 -t 10 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:06:18.336 EAL: No free 2048 kB hugepages reported on node 1 00:06:28.329 Initializing NVMe Controllers 00:06:28.329 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:06:28.329 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:06:28.329 Initialization complete. Launching workers. 00:06:28.329 ======================================================== 00:06:28.329 Latency(us) 00:06:28.329 Device Information : IOPS MiB/s Average min max 00:06:28.329 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 18209.80 71.13 3516.08 625.55 15483.54 00:06:28.329 ======================================================== 00:06:28.329 Total : 18209.80 71.13 3516.08 625.55 15483.54 00:06:28.329 00:06:28.329 17:52:21 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@65 -- # trap - SIGINT SIGTERM EXIT 00:06:28.329 17:52:21 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@66 -- # nvmftestfini 00:06:28.329 17:52:21 nvmf_tcp.nvmf_example -- nvmf/common.sh@488 -- # nvmfcleanup 00:06:28.329 17:52:21 nvmf_tcp.nvmf_example -- nvmf/common.sh@117 -- # sync 00:06:28.329 17:52:21 nvmf_tcp.nvmf_example -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:06:28.329 17:52:21 nvmf_tcp.nvmf_example -- nvmf/common.sh@120 -- # set +e 00:06:28.329 17:52:21 nvmf_tcp.nvmf_example -- nvmf/common.sh@121 -- # for i in {1..20} 00:06:28.329 17:52:21 nvmf_tcp.nvmf_example -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:06:28.329 rmmod nvme_tcp 00:06:28.329 rmmod nvme_fabrics 00:06:28.329 rmmod nvme_keyring 00:06:28.329 17:52:21 nvmf_tcp.nvmf_example -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:06:28.329 17:52:21 nvmf_tcp.nvmf_example -- nvmf/common.sh@124 -- # set -e 00:06:28.329 17:52:21 nvmf_tcp.nvmf_example -- nvmf/common.sh@125 -- # return 0 00:06:28.329 17:52:21 nvmf_tcp.nvmf_example -- nvmf/common.sh@489 -- # '[' -n 429488 ']' 00:06:28.329 17:52:21 nvmf_tcp.nvmf_example -- nvmf/common.sh@490 -- # killprocess 429488 00:06:28.329 17:52:21 nvmf_tcp.nvmf_example -- common/autotest_common.sh@948 -- # '[' -z 429488 ']' 00:06:28.329 17:52:21 nvmf_tcp.nvmf_example -- common/autotest_common.sh@952 -- # kill -0 429488 00:06:28.329 17:52:21 nvmf_tcp.nvmf_example -- common/autotest_common.sh@953 -- # uname 00:06:28.329 17:52:21 nvmf_tcp.nvmf_example -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:06:28.329 17:52:21 nvmf_tcp.nvmf_example -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 429488 00:06:28.329 17:52:22 nvmf_tcp.nvmf_example -- common/autotest_common.sh@954 -- # process_name=nvmf 00:06:28.329 17:52:22 nvmf_tcp.nvmf_example -- common/autotest_common.sh@958 -- # '[' nvmf = sudo ']' 00:06:28.329 17:52:22 nvmf_tcp.nvmf_example -- common/autotest_common.sh@966 -- # echo 'killing process with pid 429488' 00:06:28.329 killing process with pid 429488 00:06:28.329 17:52:22 nvmf_tcp.nvmf_example -- common/autotest_common.sh@967 -- # kill 429488 00:06:28.329 17:52:22 nvmf_tcp.nvmf_example -- common/autotest_common.sh@972 -- # wait 429488 00:06:28.588 nvmf threads initialize successfully 00:06:28.588 bdev subsystem init successfully 00:06:28.588 created a nvmf target service 00:06:28.588 create targets's poll groups done 00:06:28.588 all subsystems of target started 00:06:28.588 nvmf target is running 00:06:28.588 all subsystems of target stopped 00:06:28.588 destroy targets's poll groups done 00:06:28.588 destroyed the nvmf target service 00:06:28.588 bdev subsystem finish successfully 00:06:28.588 nvmf threads destroy successfully 00:06:28.588 17:52:22 nvmf_tcp.nvmf_example -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:06:28.588 17:52:22 nvmf_tcp.nvmf_example -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:06:28.588 17:52:22 nvmf_tcp.nvmf_example -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:06:28.588 17:52:22 nvmf_tcp.nvmf_example -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:06:28.588 17:52:22 nvmf_tcp.nvmf_example -- nvmf/common.sh@278 -- # remove_spdk_ns 00:06:28.588 17:52:22 nvmf_tcp.nvmf_example -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:06:28.588 17:52:22 nvmf_tcp.nvmf_example -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:06:28.588 17:52:22 nvmf_tcp.nvmf_example -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:06:31.124 17:52:24 nvmf_tcp.nvmf_example -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:06:31.124 17:52:24 nvmf_tcp.nvmf_example -- target/nvmf_example.sh@67 -- # timing_exit nvmf_example_test 00:06:31.124 17:52:24 nvmf_tcp.nvmf_example -- common/autotest_common.sh@728 -- # xtrace_disable 00:06:31.124 17:52:24 nvmf_tcp.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:06:31.124 00:06:31.124 real 0m18.704s 00:06:31.124 user 0m45.446s 00:06:31.124 sys 0m5.168s 00:06:31.124 17:52:24 nvmf_tcp.nvmf_example -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:31.124 17:52:24 nvmf_tcp.nvmf_example -- common/autotest_common.sh@10 -- # set +x 00:06:31.124 ************************************ 00:06:31.124 END TEST nvmf_example 00:06:31.124 ************************************ 00:06:31.124 17:52:24 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:06:31.124 17:52:24 nvmf_tcp -- nvmf/nvmf.sh@24 -- # run_test nvmf_filesystem /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/filesystem.sh --transport=tcp 00:06:31.124 17:52:24 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:06:31.124 17:52:24 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:31.124 17:52:24 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:06:31.124 ************************************ 00:06:31.124 START TEST nvmf_filesystem 00:06:31.124 ************************************ 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/filesystem.sh --transport=tcp 00:06:31.124 * Looking for test storage... 00:06:31.124 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- target/filesystem.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@7 -- # rpc_py=rpc_cmd 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@34 -- # set -e 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@35 -- # shopt -s nullglob 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@36 -- # shopt -s extglob 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@37 -- # shopt -s inherit_errexit 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@39 -- # '[' -z /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output ']' 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@44 -- # [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/build_config.sh ]] 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/build_config.sh 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@1 -- # CONFIG_WPDK_DIR= 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@2 -- # CONFIG_ASAN=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@3 -- # CONFIG_VBDEV_COMPRESS=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@4 -- # CONFIG_HAVE_EXECINFO_H=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@5 -- # CONFIG_USDT=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@6 -- # CONFIG_CUSTOMOCF=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@7 -- # CONFIG_PREFIX=/usr/local 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@8 -- # CONFIG_RBD=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@9 -- # CONFIG_LIBDIR= 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@10 -- # CONFIG_IDXD=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@11 -- # CONFIG_NVME_CUSE=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@12 -- # CONFIG_SMA=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@13 -- # CONFIG_VTUNE=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@14 -- # CONFIG_TSAN=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@15 -- # CONFIG_RDMA_SEND_WITH_INVAL=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@16 -- # CONFIG_VFIO_USER_DIR= 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@17 -- # CONFIG_PGO_CAPTURE=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@18 -- # CONFIG_HAVE_UUID_GENERATE_SHA1=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@19 -- # CONFIG_ENV=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@20 -- # CONFIG_LTO=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@21 -- # CONFIG_ISCSI_INITIATOR=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@22 -- # CONFIG_CET=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@23 -- # CONFIG_VBDEV_COMPRESS_MLX5=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@24 -- # CONFIG_OCF_PATH= 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@25 -- # CONFIG_RDMA_SET_TOS=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@26 -- # CONFIG_HAVE_ARC4RANDOM=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@27 -- # CONFIG_HAVE_LIBARCHIVE=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@28 -- # CONFIG_UBLK=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@29 -- # CONFIG_ISAL_CRYPTO=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@30 -- # CONFIG_OPENSSL_PATH= 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@31 -- # CONFIG_OCF=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@32 -- # CONFIG_FUSE=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@33 -- # CONFIG_VTUNE_DIR= 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@34 -- # CONFIG_FUZZER_LIB= 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@35 -- # CONFIG_FUZZER=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@36 -- # CONFIG_DPDK_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@37 -- # CONFIG_CRYPTO=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@38 -- # CONFIG_PGO_USE=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@39 -- # CONFIG_VHOST=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@40 -- # CONFIG_DAOS=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@41 -- # CONFIG_DPDK_INC_DIR= 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@42 -- # CONFIG_DAOS_DIR= 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@43 -- # CONFIG_UNIT_TESTS=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@44 -- # CONFIG_RDMA_SET_ACK_TIMEOUT=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@45 -- # CONFIG_VIRTIO=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@46 -- # CONFIG_DPDK_UADK=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@47 -- # CONFIG_COVERAGE=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@48 -- # CONFIG_RDMA=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@49 -- # CONFIG_FIO_SOURCE_DIR=/usr/src/fio 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@50 -- # CONFIG_URING_PATH= 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@51 -- # CONFIG_XNVME=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@52 -- # CONFIG_VFIO_USER=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@53 -- # CONFIG_ARCH=native 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@54 -- # CONFIG_HAVE_EVP_MAC=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@55 -- # CONFIG_URING_ZNS=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@56 -- # CONFIG_WERROR=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@57 -- # CONFIG_HAVE_LIBBSD=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@58 -- # CONFIG_UBSAN=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@59 -- # CONFIG_IPSEC_MB_DIR= 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@60 -- # CONFIG_GOLANG=n 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@61 -- # CONFIG_ISAL=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@62 -- # CONFIG_IDXD_KERNEL=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@63 -- # CONFIG_DPDK_LIB_DIR= 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@64 -- # CONFIG_RDMA_PROV=verbs 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@65 -- # CONFIG_APPS=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@66 -- # CONFIG_SHARED=y 00:06:31.124 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@67 -- # CONFIG_HAVE_KEYUTILS=y 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@68 -- # CONFIG_FC_PATH= 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@69 -- # CONFIG_DPDK_PKG_CONFIG=n 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@70 -- # CONFIG_FC=n 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@71 -- # CONFIG_AVAHI=n 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@72 -- # CONFIG_FIO_PLUGIN=y 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@73 -- # CONFIG_RAID5F=n 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@74 -- # CONFIG_EXAMPLES=y 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@75 -- # CONFIG_TESTS=y 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@76 -- # CONFIG_CRYPTO_MLX5=n 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@77 -- # CONFIG_MAX_LCORES=128 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@78 -- # CONFIG_IPSEC_MB=n 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@79 -- # CONFIG_PGO_DIR= 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@80 -- # CONFIG_DEBUG=y 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@81 -- # CONFIG_DPDK_COMPRESSDEV=n 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@82 -- # CONFIG_CROSS_PREFIX= 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/build_config.sh@83 -- # CONFIG_URING=n 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@54 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/applications.sh 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/applications.sh@8 -- # dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/applications.sh 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/applications.sh@8 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/applications.sh@8 -- # _root=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/applications.sh@9 -- # _root=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/applications.sh@10 -- # _app_dir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/applications.sh@11 -- # _test_app_dir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/applications.sh@12 -- # _examples_dir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/applications.sh@14 -- # VHOST_FUZZ_APP=("$_test_app_dir/fuzz/vhost_fuzz/vhost_fuzz") 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/applications.sh@15 -- # ISCSI_APP=("$_app_dir/iscsi_tgt") 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/applications.sh@16 -- # NVMF_APP=("$_app_dir/nvmf_tgt") 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/applications.sh@17 -- # VHOST_APP=("$_app_dir/vhost") 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/applications.sh@18 -- # DD_APP=("$_app_dir/spdk_dd") 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/applications.sh@19 -- # SPDK_APP=("$_app_dir/spdk_tgt") 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/applications.sh@22 -- # [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk/config.h ]] 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/applications.sh@23 -- # [[ #ifndef SPDK_CONFIG_H 00:06:31.125 #define SPDK_CONFIG_H 00:06:31.125 #define SPDK_CONFIG_APPS 1 00:06:31.125 #define SPDK_CONFIG_ARCH native 00:06:31.125 #undef SPDK_CONFIG_ASAN 00:06:31.125 #undef SPDK_CONFIG_AVAHI 00:06:31.125 #undef SPDK_CONFIG_CET 00:06:31.125 #define SPDK_CONFIG_COVERAGE 1 00:06:31.125 #define SPDK_CONFIG_CROSS_PREFIX 00:06:31.125 #undef SPDK_CONFIG_CRYPTO 00:06:31.125 #undef SPDK_CONFIG_CRYPTO_MLX5 00:06:31.125 #undef SPDK_CONFIG_CUSTOMOCF 00:06:31.125 #undef SPDK_CONFIG_DAOS 00:06:31.125 #define SPDK_CONFIG_DAOS_DIR 00:06:31.125 #define SPDK_CONFIG_DEBUG 1 00:06:31.125 #undef SPDK_CONFIG_DPDK_COMPRESSDEV 00:06:31.125 #define SPDK_CONFIG_DPDK_DIR /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build 00:06:31.125 #define SPDK_CONFIG_DPDK_INC_DIR 00:06:31.125 #define SPDK_CONFIG_DPDK_LIB_DIR 00:06:31.125 #undef SPDK_CONFIG_DPDK_PKG_CONFIG 00:06:31.125 #undef SPDK_CONFIG_DPDK_UADK 00:06:31.125 #define SPDK_CONFIG_ENV /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/lib/env_dpdk 00:06:31.125 #define SPDK_CONFIG_EXAMPLES 1 00:06:31.125 #undef SPDK_CONFIG_FC 00:06:31.125 #define SPDK_CONFIG_FC_PATH 00:06:31.125 #define SPDK_CONFIG_FIO_PLUGIN 1 00:06:31.125 #define SPDK_CONFIG_FIO_SOURCE_DIR /usr/src/fio 00:06:31.125 #undef SPDK_CONFIG_FUSE 00:06:31.125 #undef SPDK_CONFIG_FUZZER 00:06:31.125 #define SPDK_CONFIG_FUZZER_LIB 00:06:31.125 #undef SPDK_CONFIG_GOLANG 00:06:31.125 #define SPDK_CONFIG_HAVE_ARC4RANDOM 1 00:06:31.125 #define SPDK_CONFIG_HAVE_EVP_MAC 1 00:06:31.125 #define SPDK_CONFIG_HAVE_EXECINFO_H 1 00:06:31.125 #define SPDK_CONFIG_HAVE_KEYUTILS 1 00:06:31.125 #undef SPDK_CONFIG_HAVE_LIBARCHIVE 00:06:31.125 #undef SPDK_CONFIG_HAVE_LIBBSD 00:06:31.125 #define SPDK_CONFIG_HAVE_UUID_GENERATE_SHA1 1 00:06:31.125 #define SPDK_CONFIG_IDXD 1 00:06:31.125 #define SPDK_CONFIG_IDXD_KERNEL 1 00:06:31.125 #undef SPDK_CONFIG_IPSEC_MB 00:06:31.125 #define SPDK_CONFIG_IPSEC_MB_DIR 00:06:31.125 #define SPDK_CONFIG_ISAL 1 00:06:31.125 #define SPDK_CONFIG_ISAL_CRYPTO 1 00:06:31.125 #define SPDK_CONFIG_ISCSI_INITIATOR 1 00:06:31.125 #define SPDK_CONFIG_LIBDIR 00:06:31.125 #undef SPDK_CONFIG_LTO 00:06:31.125 #define SPDK_CONFIG_MAX_LCORES 128 00:06:31.125 #define SPDK_CONFIG_NVME_CUSE 1 00:06:31.125 #undef SPDK_CONFIG_OCF 00:06:31.125 #define SPDK_CONFIG_OCF_PATH 00:06:31.125 #define SPDK_CONFIG_OPENSSL_PATH 00:06:31.125 #undef SPDK_CONFIG_PGO_CAPTURE 00:06:31.125 #define SPDK_CONFIG_PGO_DIR 00:06:31.125 #undef SPDK_CONFIG_PGO_USE 00:06:31.125 #define SPDK_CONFIG_PREFIX /usr/local 00:06:31.125 #undef SPDK_CONFIG_RAID5F 00:06:31.125 #undef SPDK_CONFIG_RBD 00:06:31.125 #define SPDK_CONFIG_RDMA 1 00:06:31.125 #define SPDK_CONFIG_RDMA_PROV verbs 00:06:31.125 #define SPDK_CONFIG_RDMA_SEND_WITH_INVAL 1 00:06:31.125 #define SPDK_CONFIG_RDMA_SET_ACK_TIMEOUT 1 00:06:31.125 #define SPDK_CONFIG_RDMA_SET_TOS 1 00:06:31.125 #define SPDK_CONFIG_SHARED 1 00:06:31.125 #undef SPDK_CONFIG_SMA 00:06:31.125 #define SPDK_CONFIG_TESTS 1 00:06:31.125 #undef SPDK_CONFIG_TSAN 00:06:31.125 #define SPDK_CONFIG_UBLK 1 00:06:31.125 #define SPDK_CONFIG_UBSAN 1 00:06:31.125 #undef SPDK_CONFIG_UNIT_TESTS 00:06:31.125 #undef SPDK_CONFIG_URING 00:06:31.125 #define SPDK_CONFIG_URING_PATH 00:06:31.125 #undef SPDK_CONFIG_URING_ZNS 00:06:31.125 #undef SPDK_CONFIG_USDT 00:06:31.125 #undef SPDK_CONFIG_VBDEV_COMPRESS 00:06:31.125 #undef SPDK_CONFIG_VBDEV_COMPRESS_MLX5 00:06:31.125 #define SPDK_CONFIG_VFIO_USER 1 00:06:31.125 #define SPDK_CONFIG_VFIO_USER_DIR 00:06:31.125 #define SPDK_CONFIG_VHOST 1 00:06:31.125 #define SPDK_CONFIG_VIRTIO 1 00:06:31.125 #undef SPDK_CONFIG_VTUNE 00:06:31.125 #define SPDK_CONFIG_VTUNE_DIR 00:06:31.125 #define SPDK_CONFIG_WERROR 1 00:06:31.125 #define SPDK_CONFIG_WPDK_DIR 00:06:31.125 #undef SPDK_CONFIG_XNVME 00:06:31.125 #endif /* SPDK_CONFIG_H */ == *\#\d\e\f\i\n\e\ \S\P\D\K\_\C\O\N\F\I\G\_\D\E\B\U\G* ]] 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/applications.sh@24 -- # (( SPDK_AUTOTEST_DEBUG_APPS )) 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@55 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- paths/export.sh@5 -- # export PATH 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@56 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/common 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@6 -- # dirname /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/common 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@6 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@6 -- # _pmdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@7 -- # readlink -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/../../../ 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@7 -- # _pmrootdir=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@64 -- # TEST_TAG=N/A 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@65 -- # TEST_TAG_FILE=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/.run_test_name 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@67 -- # PM_OUTPUTDIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@68 -- # uname -s 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@68 -- # PM_OS=Linux 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@70 -- # MONITOR_RESOURCES_SUDO=() 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@70 -- # declare -A MONITOR_RESOURCES_SUDO 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@71 -- # MONITOR_RESOURCES_SUDO["collect-bmc-pm"]=1 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@72 -- # MONITOR_RESOURCES_SUDO["collect-cpu-load"]=0 00:06:31.125 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@73 -- # MONITOR_RESOURCES_SUDO["collect-cpu-temp"]=0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@74 -- # MONITOR_RESOURCES_SUDO["collect-vmstat"]=0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@76 -- # SUDO[0]= 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@76 -- # SUDO[1]='sudo -E' 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@78 -- # MONITOR_RESOURCES=(collect-cpu-load collect-vmstat) 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@79 -- # [[ Linux == FreeBSD ]] 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@81 -- # [[ Linux == Linux ]] 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@81 -- # [[ ............................... != QEMU ]] 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@81 -- # [[ ! -e /.dockerenv ]] 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@84 -- # MONITOR_RESOURCES+=(collect-cpu-temp) 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@85 -- # MONITOR_RESOURCES+=(collect-bmc-pm) 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- pm/common@88 -- # [[ ! -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power ]] 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@58 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@59 -- # export RUN_NIGHTLY 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@62 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@63 -- # export SPDK_AUTOTEST_DEBUG_APPS 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@64 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@65 -- # export SPDK_RUN_VALGRIND 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@66 -- # : 1 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@67 -- # export SPDK_RUN_FUNCTIONAL_TEST 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@68 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@69 -- # export SPDK_TEST_UNITTEST 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@70 -- # : 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@71 -- # export SPDK_TEST_AUTOBUILD 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@72 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@73 -- # export SPDK_TEST_RELEASE_BUILD 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@74 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@75 -- # export SPDK_TEST_ISAL 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@76 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@77 -- # export SPDK_TEST_ISCSI 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@78 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@79 -- # export SPDK_TEST_ISCSI_INITIATOR 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@80 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@81 -- # export SPDK_TEST_NVME 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@82 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@83 -- # export SPDK_TEST_NVME_PMR 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@84 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@85 -- # export SPDK_TEST_NVME_BP 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@86 -- # : 1 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@87 -- # export SPDK_TEST_NVME_CLI 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@88 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@89 -- # export SPDK_TEST_NVME_CUSE 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@90 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@91 -- # export SPDK_TEST_NVME_FDP 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@92 -- # : 1 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@93 -- # export SPDK_TEST_NVMF 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@94 -- # : 1 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@95 -- # export SPDK_TEST_VFIOUSER 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@96 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@97 -- # export SPDK_TEST_VFIOUSER_QEMU 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@98 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@99 -- # export SPDK_TEST_FUZZER 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@100 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@101 -- # export SPDK_TEST_FUZZER_SHORT 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@102 -- # : tcp 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@103 -- # export SPDK_TEST_NVMF_TRANSPORT 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@104 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@105 -- # export SPDK_TEST_RBD 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@106 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@107 -- # export SPDK_TEST_VHOST 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@108 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@109 -- # export SPDK_TEST_BLOCKDEV 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@110 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@111 -- # export SPDK_TEST_IOAT 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@112 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@113 -- # export SPDK_TEST_BLOBFS 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@114 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@115 -- # export SPDK_TEST_VHOST_INIT 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@116 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@117 -- # export SPDK_TEST_LVOL 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@118 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@119 -- # export SPDK_TEST_VBDEV_COMPRESS 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@120 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@121 -- # export SPDK_RUN_ASAN 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@122 -- # : 1 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@123 -- # export SPDK_RUN_UBSAN 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@124 -- # : 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@125 -- # export SPDK_RUN_EXTERNAL_DPDK 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@126 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@127 -- # export SPDK_RUN_NON_ROOT 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@128 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@129 -- # export SPDK_TEST_CRYPTO 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@130 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@131 -- # export SPDK_TEST_FTL 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@132 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@133 -- # export SPDK_TEST_OCF 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@134 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@135 -- # export SPDK_TEST_VMD 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@136 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@137 -- # export SPDK_TEST_OPAL 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@138 -- # : 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@139 -- # export SPDK_TEST_NATIVE_DPDK 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@140 -- # : true 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@141 -- # export SPDK_AUTOTEST_X 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@142 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@143 -- # export SPDK_TEST_RAID5 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@144 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@145 -- # export SPDK_TEST_URING 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@146 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@147 -- # export SPDK_TEST_USDT 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@148 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@149 -- # export SPDK_TEST_USE_IGB_UIO 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@150 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@151 -- # export SPDK_TEST_SCHEDULER 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@152 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@153 -- # export SPDK_TEST_SCANBUILD 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@154 -- # : e810 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@155 -- # export SPDK_TEST_NVMF_NICS 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@156 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@157 -- # export SPDK_TEST_SMA 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@158 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@159 -- # export SPDK_TEST_DAOS 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@160 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@161 -- # export SPDK_TEST_XNVME 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@162 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@163 -- # export SPDK_TEST_ACCEL_DSA 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@164 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@165 -- # export SPDK_TEST_ACCEL_IAA 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@167 -- # : 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@168 -- # export SPDK_TEST_FUZZER_TARGET 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@169 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@170 -- # export SPDK_TEST_NVMF_MDNS 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@171 -- # : 0 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@172 -- # export SPDK_JSONRPC_GO_CLIENT 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@175 -- # export SPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@175 -- # SPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@176 -- # export DPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@176 -- # DPDK_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@177 -- # export VFIO_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:06:31.126 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@177 -- # VFIO_LIB_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@178 -- # export LD_LIBRARY_PATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@178 -- # LD_LIBRARY_PATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/build/lib:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/libvfio-user/usr/local/lib 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@181 -- # export PCI_BLOCK_SYNC_ON_RESET=yes 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@181 -- # PCI_BLOCK_SYNC_ON_RESET=yes 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@185 -- # export PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@185 -- # PYTHONPATH=:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/rpc_plugins:/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/python 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@189 -- # export PYTHONDONTWRITEBYTECODE=1 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@189 -- # PYTHONDONTWRITEBYTECODE=1 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@193 -- # export ASAN_OPTIONS=new_delete_type_mismatch=0:disable_coredump=0:abort_on_error=1:use_sigaltstack=0 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@193 -- # ASAN_OPTIONS=new_delete_type_mismatch=0:disable_coredump=0:abort_on_error=1:use_sigaltstack=0 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@194 -- # export UBSAN_OPTIONS=halt_on_error=1:print_stacktrace=1:abort_on_error=1:disable_coredump=0:exitcode=134 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@194 -- # UBSAN_OPTIONS=halt_on_error=1:print_stacktrace=1:abort_on_error=1:disable_coredump=0:exitcode=134 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@198 -- # asan_suppression_file=/var/tmp/asan_suppression_file 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@199 -- # rm -rf /var/tmp/asan_suppression_file 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@200 -- # cat 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@236 -- # echo leak:libfuse3.so 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@238 -- # export LSAN_OPTIONS=suppressions=/var/tmp/asan_suppression_file 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@238 -- # LSAN_OPTIONS=suppressions=/var/tmp/asan_suppression_file 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@240 -- # export DEFAULT_RPC_ADDR=/var/tmp/spdk.sock 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@240 -- # DEFAULT_RPC_ADDR=/var/tmp/spdk.sock 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@242 -- # '[' -z /var/spdk/dependencies ']' 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@245 -- # export DEPENDENCY_DIR 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@249 -- # export SPDK_BIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@249 -- # SPDK_BIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@250 -- # export SPDK_EXAMPLE_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@250 -- # SPDK_EXAMPLE_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@253 -- # export QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@253 -- # QEMU_BIN=/usr/local/qemu/vanilla-latest/bin/qemu-system-x86_64 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@254 -- # export VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@254 -- # VFIO_QEMU_BIN=/usr/local/qemu/vfio-user-latest/bin/qemu-system-x86_64 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@256 -- # export AR_TOOL=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/ar-xnvme-fixer 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@256 -- # AR_TOOL=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/ar-xnvme-fixer 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@259 -- # export UNBIND_ENTIRE_IOMMU_GROUP=yes 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@259 -- # UNBIND_ENTIRE_IOMMU_GROUP=yes 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@262 -- # '[' 0 -eq 0 ']' 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@263 -- # export valgrind= 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@263 -- # valgrind= 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@269 -- # uname -s 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@269 -- # '[' Linux = Linux ']' 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@270 -- # HUGEMEM=4096 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@271 -- # export CLEAR_HUGE=yes 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@271 -- # CLEAR_HUGE=yes 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@272 -- # [[ 0 -eq 1 ]] 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@272 -- # [[ 0 -eq 1 ]] 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@279 -- # MAKE=make 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@280 -- # MAKEFLAGS=-j96 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@296 -- # export HUGEMEM=4096 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@296 -- # HUGEMEM=4096 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@298 -- # NO_HUGE=() 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@299 -- # TEST_MODE= 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@300 -- # for i in "$@" 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@301 -- # case "$i" in 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@306 -- # TEST_TRANSPORT=tcp 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@318 -- # [[ -z 431898 ]] 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@318 -- # kill -0 431898 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@1680 -- # set_test_storage 2147483648 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@328 -- # [[ -v testdir ]] 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@330 -- # local requested_size=2147483648 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@331 -- # local mount target_dir 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@333 -- # local -A mounts fss sizes avails uses 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@334 -- # local source fs size avail mount use 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@336 -- # local storage_fallback storage_candidates 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@338 -- # mktemp -udt spdk.XXXXXX 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@338 -- # storage_fallback=/tmp/spdk.5Ot9Ek 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@343 -- # storage_candidates=("$testdir" "$storage_fallback/tests/${testdir##*/}" "$storage_fallback") 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@345 -- # [[ -n '' ]] 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@350 -- # [[ -n '' ]] 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@355 -- # mkdir -p /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target /tmp/spdk.5Ot9Ek/tests/target /tmp/spdk.5Ot9Ek 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@358 -- # requested_size=2214592512 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@327 -- # df -T 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@327 -- # grep -v Filesystem 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@361 -- # mounts["$mount"]=spdk_devtmpfs 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@361 -- # fss["$mount"]=devtmpfs 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@362 -- # avails["$mount"]=67108864 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@362 -- # sizes["$mount"]=67108864 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@363 -- # uses["$mount"]=0 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@361 -- # mounts["$mount"]=/dev/pmem0 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@361 -- # fss["$mount"]=ext2 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@362 -- # avails["$mount"]=950202368 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@362 -- # sizes["$mount"]=5284429824 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@363 -- # uses["$mount"]=4334227456 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@361 -- # mounts["$mount"]=spdk_root 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@361 -- # fss["$mount"]=overlay 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@362 -- # avails["$mount"]=189612310528 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@362 -- # sizes["$mount"]=195974299648 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@363 -- # uses["$mount"]=6361989120 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@361 -- # mounts["$mount"]=tmpfs 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@361 -- # fss["$mount"]=tmpfs 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@362 -- # avails["$mount"]=97983774720 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@362 -- # sizes["$mount"]=97987149824 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@363 -- # uses["$mount"]=3375104 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@361 -- # mounts["$mount"]=tmpfs 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@361 -- # fss["$mount"]=tmpfs 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@362 -- # avails["$mount"]=39185485824 00:06:31.127 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@362 -- # sizes["$mount"]=39194861568 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@363 -- # uses["$mount"]=9375744 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@361 -- # mounts["$mount"]=tmpfs 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@361 -- # fss["$mount"]=tmpfs 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@362 -- # avails["$mount"]=97986420736 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@362 -- # sizes["$mount"]=97987149824 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@363 -- # uses["$mount"]=729088 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@361 -- # mounts["$mount"]=tmpfs 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@361 -- # fss["$mount"]=tmpfs 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@362 -- # avails["$mount"]=19597422592 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@362 -- # sizes["$mount"]=19597426688 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@363 -- # uses["$mount"]=4096 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@360 -- # read -r source fs size use avail _ mount 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@366 -- # printf '* Looking for test storage...\n' 00:06:31.128 * Looking for test storage... 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@368 -- # local target_space new_size 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@369 -- # for target_dir in "${storage_candidates[@]}" 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@372 -- # df /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@372 -- # awk '$1 !~ /Filesystem/{print $6}' 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@372 -- # mount=/ 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@374 -- # target_space=189612310528 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@375 -- # (( target_space == 0 || target_space < requested_size )) 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@378 -- # (( target_space >= requested_size )) 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@380 -- # [[ overlay == tmpfs ]] 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@380 -- # [[ overlay == ramfs ]] 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@380 -- # [[ / == / ]] 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@381 -- # new_size=8576581632 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@382 -- # (( new_size * 100 / sizes[/] > 95 )) 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@387 -- # export SPDK_TEST_STORAGE=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@387 -- # SPDK_TEST_STORAGE=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@388 -- # printf '* Found test storage at %s\n' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:06:31.128 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@389 -- # return 0 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@1682 -- # set -o errtrace 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@1683 -- # shopt -s extdebug 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@1684 -- # trap 'trap - ERR; print_backtrace >&2' ERR 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@1686 -- # PS4=' \t ${test_domain:-} -- ${BASH_SOURCE#${BASH_SOURCE%/*/*}/}@${LINENO} -- \$ ' 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@1687 -- # true 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@1689 -- # xtrace_fd 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@25 -- # [[ -n 14 ]] 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@25 -- # [[ -e /proc/self/fd/14 ]] 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@27 -- # exec 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@29 -- # exec 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@31 -- # xtrace_restore 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@16 -- # unset -v 'X_STACK[0 - 1 < 0 ? 0 : 0 - 1]' 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@17 -- # (( 0 == 0 )) 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@18 -- # set -x 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- target/filesystem.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@7 -- # uname -s 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- paths/export.sh@5 -- # export PATH 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@47 -- # : 0 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:06:31.128 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:06:31.129 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@51 -- # have_pci_nics=0 00:06:31.129 17:52:24 nvmf_tcp.nvmf_filesystem -- target/filesystem.sh@12 -- # MALLOC_BDEV_SIZE=512 00:06:31.129 17:52:24 nvmf_tcp.nvmf_filesystem -- target/filesystem.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:06:31.129 17:52:24 nvmf_tcp.nvmf_filesystem -- target/filesystem.sh@15 -- # nvmftestinit 00:06:31.129 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:06:31.129 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:06:31.129 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@448 -- # prepare_net_devs 00:06:31.129 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@410 -- # local -g is_hw=no 00:06:31.129 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@412 -- # remove_spdk_ns 00:06:31.129 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:06:31.129 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:06:31.129 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:06:31.129 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:06:31.129 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:06:31.129 17:52:24 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@285 -- # xtrace_disable 00:06:31.129 17:52:24 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@10 -- # set +x 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@291 -- # pci_devs=() 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@291 -- # local -a pci_devs 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@292 -- # pci_net_devs=() 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@293 -- # pci_drivers=() 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@293 -- # local -A pci_drivers 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@295 -- # net_devs=() 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@295 -- # local -ga net_devs 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@296 -- # e810=() 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@296 -- # local -ga e810 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@297 -- # x722=() 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@297 -- # local -ga x722 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@298 -- # mlx=() 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@298 -- # local -ga mlx 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:06:36.408 Found 0000:86:00.0 (0x8086 - 0x159b) 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:06:36.408 Found 0000:86:00.1 (0x8086 - 0x159b) 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:06:36.408 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@390 -- # [[ up == up ]] 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:06:36.409 Found net devices under 0000:86:00.0: cvl_0_0 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@390 -- # [[ up == up ]] 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:06:36.409 Found net devices under 0000:86:00.1: cvl_0_1 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@414 -- # is_hw=yes 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:06:36.409 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:06:36.409 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.160 ms 00:06:36.409 00:06:36.409 --- 10.0.0.2 ping statistics --- 00:06:36.409 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:06:36.409 rtt min/avg/max/mdev = 0.160/0.160/0.160/0.000 ms 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:06:36.409 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:06:36.409 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.098 ms 00:06:36.409 00:06:36.409 --- 10.0.0.1 ping statistics --- 00:06:36.409 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:06:36.409 rtt min/avg/max/mdev = 0.098/0.098/0.098/0.000 ms 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@422 -- # return 0 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- target/filesystem.sh@105 -- # run_test nvmf_filesystem_no_in_capsule nvmf_filesystem_part 0 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@10 -- # set +x 00:06:36.409 ************************************ 00:06:36.409 START TEST nvmf_filesystem_no_in_capsule 00:06:36.409 ************************************ 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1123 -- # nvmf_filesystem_part 0 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@47 -- # in_capsule=0 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@49 -- # nvmfappstart -m 0xF 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@722 -- # xtrace_disable 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@481 -- # nvmfpid=434894 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@482 -- # waitforlisten 434894 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@829 -- # '[' -z 434894 ']' 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@834 -- # local max_retries=100 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:36.409 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@838 -- # xtrace_disable 00:06:36.409 17:52:29 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:36.409 [2024-07-15 17:52:29.601343] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:06:36.409 [2024-07-15 17:52:29.601383] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:36.409 EAL: No free 2048 kB hugepages reported on node 1 00:06:36.409 [2024-07-15 17:52:29.657950] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:06:36.409 [2024-07-15 17:52:29.740680] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:06:36.409 [2024-07-15 17:52:29.740720] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:06:36.409 [2024-07-15 17:52:29.740727] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:06:36.409 [2024-07-15 17:52:29.740734] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:06:36.409 [2024-07-15 17:52:29.740739] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:06:36.409 [2024-07-15 17:52:29.740793] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:06:36.409 [2024-07-15 17:52:29.740810] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:06:36.409 [2024-07-15 17:52:29.740899] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:06:36.409 [2024-07-15 17:52:29.740900] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@862 -- # return 0 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@728 -- # xtrace_disable 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@50 -- # malloc_name=Malloc1 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@52 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -c 0 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:36.978 [2024-07-15 17:52:30.438027] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@53 -- # rpc_cmd bdev_malloc_create 512 512 -b Malloc1 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:36.978 Malloc1 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@54 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@55 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@56 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:36.978 [2024-07-15 17:52:30.587294] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@58 -- # get_bdev_size Malloc1 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1378 -- # local bdev_name=Malloc1 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1379 -- # local bdev_info 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1380 -- # local bs 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1381 -- # local nb 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1382 -- # rpc_cmd bdev_get_bdevs -b Malloc1 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:36.978 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1382 -- # bdev_info='[ 00:06:36.978 { 00:06:36.978 "name": "Malloc1", 00:06:36.978 "aliases": [ 00:06:36.978 "e8f59bde-4988-4430-b1d0-2479b784583e" 00:06:36.978 ], 00:06:36.978 "product_name": "Malloc disk", 00:06:36.978 "block_size": 512, 00:06:36.978 "num_blocks": 1048576, 00:06:36.978 "uuid": "e8f59bde-4988-4430-b1d0-2479b784583e", 00:06:36.978 "assigned_rate_limits": { 00:06:36.978 "rw_ios_per_sec": 0, 00:06:36.978 "rw_mbytes_per_sec": 0, 00:06:36.978 "r_mbytes_per_sec": 0, 00:06:36.978 "w_mbytes_per_sec": 0 00:06:36.978 }, 00:06:36.978 "claimed": true, 00:06:36.978 "claim_type": "exclusive_write", 00:06:36.978 "zoned": false, 00:06:36.978 "supported_io_types": { 00:06:36.978 "read": true, 00:06:36.978 "write": true, 00:06:36.978 "unmap": true, 00:06:36.978 "flush": true, 00:06:36.978 "reset": true, 00:06:36.978 "nvme_admin": false, 00:06:36.978 "nvme_io": false, 00:06:36.978 "nvme_io_md": false, 00:06:36.978 "write_zeroes": true, 00:06:36.978 "zcopy": true, 00:06:36.978 "get_zone_info": false, 00:06:36.978 "zone_management": false, 00:06:36.978 "zone_append": false, 00:06:36.978 "compare": false, 00:06:36.978 "compare_and_write": false, 00:06:36.978 "abort": true, 00:06:36.978 "seek_hole": false, 00:06:36.978 "seek_data": false, 00:06:36.978 "copy": true, 00:06:36.978 "nvme_iov_md": false 00:06:36.978 }, 00:06:36.978 "memory_domains": [ 00:06:36.978 { 00:06:36.978 "dma_device_id": "system", 00:06:36.978 "dma_device_type": 1 00:06:36.978 }, 00:06:36.978 { 00:06:36.978 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:36.978 "dma_device_type": 2 00:06:36.978 } 00:06:36.978 ], 00:06:36.978 "driver_specific": {} 00:06:36.978 } 00:06:36.978 ]' 00:06:36.979 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1383 -- # jq '.[] .block_size' 00:06:36.979 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1383 -- # bs=512 00:06:36.979 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1384 -- # jq '.[] .num_blocks' 00:06:36.979 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1384 -- # nb=1048576 00:06:36.979 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1387 -- # bdev_size=512 00:06:36.979 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1388 -- # echo 512 00:06:36.979 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@58 -- # malloc_size=536870912 00:06:36.979 17:52:30 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@60 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:06:38.355 17:52:31 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@62 -- # waitforserial SPDKISFASTANDAWESOME 00:06:38.355 17:52:31 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1198 -- # local i=0 00:06:38.355 17:52:31 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:06:38.355 17:52:31 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:06:38.355 17:52:31 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1205 -- # sleep 2 00:06:40.261 17:52:33 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:06:40.261 17:52:33 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:06:40.261 17:52:33 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:06:40.261 17:52:33 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:06:40.261 17:52:33 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:06:40.261 17:52:33 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1208 -- # return 0 00:06:40.261 17:52:33 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@63 -- # lsblk -l -o NAME,SERIAL 00:06:40.261 17:52:33 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@63 -- # grep -oP '([\w]*)(?=\s+SPDKISFASTANDAWESOME)' 00:06:40.261 17:52:33 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@63 -- # nvme_name=nvme0n1 00:06:40.261 17:52:33 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@64 -- # sec_size_to_bytes nvme0n1 00:06:40.261 17:52:33 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- setup/common.sh@76 -- # local dev=nvme0n1 00:06:40.261 17:52:33 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- setup/common.sh@78 -- # [[ -e /sys/block/nvme0n1 ]] 00:06:40.261 17:52:33 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- setup/common.sh@80 -- # echo 536870912 00:06:40.261 17:52:33 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@64 -- # nvme_size=536870912 00:06:40.261 17:52:33 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@66 -- # mkdir -p /mnt/device 00:06:40.261 17:52:33 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@67 -- # (( nvme_size == malloc_size )) 00:06:40.261 17:52:33 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@68 -- # parted -s /dev/nvme0n1 mklabel gpt mkpart SPDK_TEST 0% 100% 00:06:40.520 17:52:33 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@69 -- # partprobe 00:06:40.778 17:52:34 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@70 -- # sleep 1 00:06:41.716 17:52:35 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@76 -- # '[' 0 -eq 0 ']' 00:06:41.716 17:52:35 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@77 -- # run_test filesystem_ext4 nvmf_filesystem_create ext4 nvme0n1 00:06:41.716 17:52:35 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:06:41.716 17:52:35 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:41.716 17:52:35 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:41.716 ************************************ 00:06:41.716 START TEST filesystem_ext4 00:06:41.716 ************************************ 00:06:41.716 17:52:35 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@1123 -- # nvmf_filesystem_create ext4 nvme0n1 00:06:41.716 17:52:35 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@18 -- # fstype=ext4 00:06:41.716 17:52:35 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:06:41.716 17:52:35 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@21 -- # make_filesystem ext4 /dev/nvme0n1p1 00:06:41.716 17:52:35 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@924 -- # local fstype=ext4 00:06:41.716 17:52:35 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@925 -- # local dev_name=/dev/nvme0n1p1 00:06:41.716 17:52:35 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@926 -- # local i=0 00:06:41.716 17:52:35 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@927 -- # local force 00:06:41.716 17:52:35 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@929 -- # '[' ext4 = ext4 ']' 00:06:41.716 17:52:35 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@930 -- # force=-F 00:06:41.716 17:52:35 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@935 -- # mkfs.ext4 -F /dev/nvme0n1p1 00:06:41.716 mke2fs 1.46.5 (30-Dec-2021) 00:06:41.975 Discarding device blocks: 0/522240 done 00:06:41.975 Creating filesystem with 522240 1k blocks and 130560 inodes 00:06:41.975 Filesystem UUID: 3ee65914-8185-473d-a99f-f14450739f04 00:06:41.975 Superblock backups stored on blocks: 00:06:41.975 8193, 24577, 40961, 57345, 73729, 204801, 221185, 401409 00:06:41.975 00:06:41.975 Allocating group tables: 0/64 done 00:06:41.975 Writing inode tables: 0/64 done 00:06:44.548 Creating journal (8192 blocks): done 00:06:44.549 Writing superblocks and filesystem accounting information: 0/64 done 00:06:44.549 00:06:44.549 17:52:37 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@943 -- # return 0 00:06:44.549 17:52:37 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:06:44.808 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:06:44.808 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@25 -- # sync 00:06:44.808 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:06:44.808 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@27 -- # sync 00:06:44.808 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@29 -- # i=0 00:06:44.808 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@30 -- # umount /mnt/device 00:06:44.808 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@37 -- # kill -0 434894 00:06:44.808 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:06:44.808 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:06:44.808 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:06:44.808 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:06:44.808 00:06:44.808 real 0m3.106s 00:06:44.808 user 0m0.031s 00:06:44.808 sys 0m0.064s 00:06:44.808 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:44.808 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_ext4 -- common/autotest_common.sh@10 -- # set +x 00:06:44.808 ************************************ 00:06:44.808 END TEST filesystem_ext4 00:06:44.808 ************************************ 00:06:45.067 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1142 -- # return 0 00:06:45.067 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@78 -- # run_test filesystem_btrfs nvmf_filesystem_create btrfs nvme0n1 00:06:45.067 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:06:45.067 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:45.067 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:45.067 ************************************ 00:06:45.067 START TEST filesystem_btrfs 00:06:45.067 ************************************ 00:06:45.067 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@1123 -- # nvmf_filesystem_create btrfs nvme0n1 00:06:45.067 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@18 -- # fstype=btrfs 00:06:45.067 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:06:45.067 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@21 -- # make_filesystem btrfs /dev/nvme0n1p1 00:06:45.067 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@924 -- # local fstype=btrfs 00:06:45.067 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@925 -- # local dev_name=/dev/nvme0n1p1 00:06:45.067 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@926 -- # local i=0 00:06:45.067 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@927 -- # local force 00:06:45.067 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@929 -- # '[' btrfs = ext4 ']' 00:06:45.067 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@932 -- # force=-f 00:06:45.067 17:52:38 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@935 -- # mkfs.btrfs -f /dev/nvme0n1p1 00:06:45.325 btrfs-progs v6.6.2 00:06:45.325 See https://btrfs.readthedocs.io for more information. 00:06:45.325 00:06:45.325 Performing full device TRIM /dev/nvme0n1p1 (510.00MiB) ... 00:06:45.325 NOTE: several default settings have changed in version 5.15, please make sure 00:06:45.325 this does not affect your deployments: 00:06:45.325 - DUP for metadata (-m dup) 00:06:45.325 - enabled no-holes (-O no-holes) 00:06:45.325 - enabled free-space-tree (-R free-space-tree) 00:06:45.325 00:06:45.325 Label: (null) 00:06:45.325 UUID: 2fc8b49e-61a0-442f-8ee0-deea68bc5945 00:06:45.325 Node size: 16384 00:06:45.325 Sector size: 4096 00:06:45.325 Filesystem size: 510.00MiB 00:06:45.325 Block group profiles: 00:06:45.325 Data: single 8.00MiB 00:06:45.325 Metadata: DUP 32.00MiB 00:06:45.326 System: DUP 8.00MiB 00:06:45.326 SSD detected: yes 00:06:45.326 Zoned device: no 00:06:45.326 Incompat features: extref, skinny-metadata, no-holes, free-space-tree 00:06:45.326 Runtime features: free-space-tree 00:06:45.326 Checksum: crc32c 00:06:45.326 Number of devices: 1 00:06:45.326 Devices: 00:06:45.326 ID SIZE PATH 00:06:45.326 1 510.00MiB /dev/nvme0n1p1 00:06:45.326 00:06:45.326 17:52:39 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@943 -- # return 0 00:06:45.326 17:52:39 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:06:46.264 17:52:39 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:06:46.264 17:52:39 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@25 -- # sync 00:06:46.264 17:52:39 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:06:46.264 17:52:39 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@27 -- # sync 00:06:46.264 17:52:39 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@29 -- # i=0 00:06:46.264 17:52:39 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@30 -- # umount /mnt/device 00:06:46.264 17:52:39 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@37 -- # kill -0 434894 00:06:46.264 17:52:39 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:06:46.264 17:52:39 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:06:46.264 17:52:39 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:06:46.264 17:52:39 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:06:46.264 00:06:46.264 real 0m1.348s 00:06:46.264 user 0m0.033s 00:06:46.264 sys 0m0.121s 00:06:46.264 17:52:39 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:46.264 17:52:39 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_btrfs -- common/autotest_common.sh@10 -- # set +x 00:06:46.264 ************************************ 00:06:46.264 END TEST filesystem_btrfs 00:06:46.264 ************************************ 00:06:46.264 17:52:39 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1142 -- # return 0 00:06:46.264 17:52:39 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@79 -- # run_test filesystem_xfs nvmf_filesystem_create xfs nvme0n1 00:06:46.265 17:52:39 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:06:46.265 17:52:39 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:46.265 17:52:39 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:46.524 ************************************ 00:06:46.524 START TEST filesystem_xfs 00:06:46.524 ************************************ 00:06:46.524 17:52:40 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@1123 -- # nvmf_filesystem_create xfs nvme0n1 00:06:46.524 17:52:40 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@18 -- # fstype=xfs 00:06:46.524 17:52:40 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:06:46.524 17:52:40 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@21 -- # make_filesystem xfs /dev/nvme0n1p1 00:06:46.524 17:52:40 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@924 -- # local fstype=xfs 00:06:46.524 17:52:40 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@925 -- # local dev_name=/dev/nvme0n1p1 00:06:46.524 17:52:40 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@926 -- # local i=0 00:06:46.524 17:52:40 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@927 -- # local force 00:06:46.524 17:52:40 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@929 -- # '[' xfs = ext4 ']' 00:06:46.524 17:52:40 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@932 -- # force=-f 00:06:46.524 17:52:40 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@935 -- # mkfs.xfs -f /dev/nvme0n1p1 00:06:46.524 meta-data=/dev/nvme0n1p1 isize=512 agcount=4, agsize=32640 blks 00:06:46.524 = sectsz=512 attr=2, projid32bit=1 00:06:46.524 = crc=1 finobt=1, sparse=1, rmapbt=0 00:06:46.524 = reflink=1 bigtime=1 inobtcount=1 nrext64=0 00:06:46.524 data = bsize=4096 blocks=130560, imaxpct=25 00:06:46.524 = sunit=0 swidth=0 blks 00:06:46.524 naming =version 2 bsize=4096 ascii-ci=0, ftype=1 00:06:46.524 log =internal log bsize=4096 blocks=16384, version=2 00:06:46.524 = sectsz=512 sunit=0 blks, lazy-count=1 00:06:46.524 realtime =none extsz=4096 blocks=0, rtextents=0 00:06:47.460 Discarding blocks...Done. 00:06:47.460 17:52:40 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@943 -- # return 0 00:06:47.460 17:52:40 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@25 -- # sync 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@27 -- # sync 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@29 -- # i=0 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@30 -- # umount /mnt/device 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@37 -- # kill -0 434894 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:06:49.989 00:06:49.989 real 0m3.396s 00:06:49.989 user 0m0.028s 00:06:49.989 sys 0m0.069s 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule.filesystem_xfs -- common/autotest_common.sh@10 -- # set +x 00:06:49.989 ************************************ 00:06:49.989 END TEST filesystem_xfs 00:06:49.989 ************************************ 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1142 -- # return 0 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@91 -- # flock /dev/nvme0n1 parted -s /dev/nvme0n1 rm 1 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@93 -- # sync 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@94 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:06:49.989 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@95 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1219 -- # local i=0 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1231 -- # return 0 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@97 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@101 -- # killprocess 434894 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@948 -- # '[' -z 434894 ']' 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@952 -- # kill -0 434894 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@953 -- # uname 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 434894 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@966 -- # echo 'killing process with pid 434894' 00:06:49.989 killing process with pid 434894 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@967 -- # kill 434894 00:06:49.989 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@972 -- # wait 434894 00:06:50.554 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- target/filesystem.sh@102 -- # nvmfpid= 00:06:50.554 00:06:50.554 real 0m14.439s 00:06:50.554 user 0m56.847s 00:06:50.554 sys 0m1.199s 00:06:50.554 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:50.554 17:52:43 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_no_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:50.554 ************************************ 00:06:50.554 END TEST nvmf_filesystem_no_in_capsule 00:06:50.554 ************************************ 00:06:50.554 17:52:44 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@1142 -- # return 0 00:06:50.554 17:52:44 nvmf_tcp.nvmf_filesystem -- target/filesystem.sh@106 -- # run_test nvmf_filesystem_in_capsule nvmf_filesystem_part 4096 00:06:50.554 17:52:44 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:06:50.554 17:52:44 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:50.554 17:52:44 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@10 -- # set +x 00:06:50.554 ************************************ 00:06:50.554 START TEST nvmf_filesystem_in_capsule 00:06:50.554 ************************************ 00:06:50.554 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1123 -- # nvmf_filesystem_part 4096 00:06:50.554 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@47 -- # in_capsule=4096 00:06:50.554 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@49 -- # nvmfappstart -m 0xF 00:06:50.554 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:06:50.554 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@722 -- # xtrace_disable 00:06:50.554 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:50.554 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@481 -- # nvmfpid=437457 00:06:50.554 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@482 -- # waitforlisten 437457 00:06:50.554 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:06:50.554 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@829 -- # '[' -z 437457 ']' 00:06:50.554 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:06:50.554 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@834 -- # local max_retries=100 00:06:50.554 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:06:50.554 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:06:50.554 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@838 -- # xtrace_disable 00:06:50.554 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:50.554 [2024-07-15 17:52:44.112546] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:06:50.554 [2024-07-15 17:52:44.112589] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:06:50.554 EAL: No free 2048 kB hugepages reported on node 1 00:06:50.554 [2024-07-15 17:52:44.170558] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:06:50.554 [2024-07-15 17:52:44.241229] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:06:50.554 [2024-07-15 17:52:44.241270] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:06:50.554 [2024-07-15 17:52:44.241277] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:06:50.554 [2024-07-15 17:52:44.241283] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:06:50.554 [2024-07-15 17:52:44.241288] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:06:50.554 [2024-07-15 17:52:44.241349] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:06:50.554 [2024-07-15 17:52:44.241441] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:06:50.554 [2024-07-15 17:52:44.241530] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:06:50.554 [2024-07-15 17:52:44.241531] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:06:51.486 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:06:51.486 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@862 -- # return 0 00:06:51.486 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:06:51.486 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@728 -- # xtrace_disable 00:06:51.486 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:51.486 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:06:51.486 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@50 -- # malloc_name=Malloc1 00:06:51.486 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@52 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -c 4096 00:06:51.486 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:51.486 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:51.486 [2024-07-15 17:52:44.965124] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:06:51.486 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:51.486 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@53 -- # rpc_cmd bdev_malloc_create 512 512 -b Malloc1 00:06:51.486 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:51.486 17:52:44 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:51.486 Malloc1 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@54 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@55 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@56 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:51.486 [2024-07-15 17:52:45.108418] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@58 -- # get_bdev_size Malloc1 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1378 -- # local bdev_name=Malloc1 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1379 -- # local bdev_info 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1380 -- # local bs 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1381 -- # local nb 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1382 -- # rpc_cmd bdev_get_bdevs -b Malloc1 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@559 -- # xtrace_disable 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:06:51.486 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1382 -- # bdev_info='[ 00:06:51.486 { 00:06:51.486 "name": "Malloc1", 00:06:51.486 "aliases": [ 00:06:51.487 "8951a388-3482-4992-956d-2d1c265e9dd5" 00:06:51.487 ], 00:06:51.487 "product_name": "Malloc disk", 00:06:51.487 "block_size": 512, 00:06:51.487 "num_blocks": 1048576, 00:06:51.487 "uuid": "8951a388-3482-4992-956d-2d1c265e9dd5", 00:06:51.487 "assigned_rate_limits": { 00:06:51.487 "rw_ios_per_sec": 0, 00:06:51.487 "rw_mbytes_per_sec": 0, 00:06:51.487 "r_mbytes_per_sec": 0, 00:06:51.487 "w_mbytes_per_sec": 0 00:06:51.487 }, 00:06:51.487 "claimed": true, 00:06:51.487 "claim_type": "exclusive_write", 00:06:51.487 "zoned": false, 00:06:51.487 "supported_io_types": { 00:06:51.487 "read": true, 00:06:51.487 "write": true, 00:06:51.487 "unmap": true, 00:06:51.487 "flush": true, 00:06:51.487 "reset": true, 00:06:51.487 "nvme_admin": false, 00:06:51.487 "nvme_io": false, 00:06:51.487 "nvme_io_md": false, 00:06:51.487 "write_zeroes": true, 00:06:51.487 "zcopy": true, 00:06:51.487 "get_zone_info": false, 00:06:51.487 "zone_management": false, 00:06:51.487 "zone_append": false, 00:06:51.487 "compare": false, 00:06:51.487 "compare_and_write": false, 00:06:51.487 "abort": true, 00:06:51.487 "seek_hole": false, 00:06:51.487 "seek_data": false, 00:06:51.487 "copy": true, 00:06:51.487 "nvme_iov_md": false 00:06:51.487 }, 00:06:51.487 "memory_domains": [ 00:06:51.487 { 00:06:51.487 "dma_device_id": "system", 00:06:51.487 "dma_device_type": 1 00:06:51.487 }, 00:06:51.487 { 00:06:51.487 "dma_device_id": "SPDK_ACCEL_DMA_DEVICE", 00:06:51.487 "dma_device_type": 2 00:06:51.487 } 00:06:51.487 ], 00:06:51.487 "driver_specific": {} 00:06:51.487 } 00:06:51.487 ]' 00:06:51.487 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1383 -- # jq '.[] .block_size' 00:06:51.487 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1383 -- # bs=512 00:06:51.487 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1384 -- # jq '.[] .num_blocks' 00:06:51.745 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1384 -- # nb=1048576 00:06:51.745 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1387 -- # bdev_size=512 00:06:51.745 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1388 -- # echo 512 00:06:51.745 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@58 -- # malloc_size=536870912 00:06:51.745 17:52:45 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@60 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:06:52.682 17:52:46 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@62 -- # waitforserial SPDKISFASTANDAWESOME 00:06:52.682 17:52:46 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1198 -- # local i=0 00:06:52.682 17:52:46 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:06:52.682 17:52:46 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:06:52.682 17:52:46 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1205 -- # sleep 2 00:06:55.217 17:52:48 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:06:55.217 17:52:48 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:06:55.217 17:52:48 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:06:55.217 17:52:48 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:06:55.217 17:52:48 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:06:55.217 17:52:48 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1208 -- # return 0 00:06:55.217 17:52:48 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@63 -- # lsblk -l -o NAME,SERIAL 00:06:55.217 17:52:48 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@63 -- # grep -oP '([\w]*)(?=\s+SPDKISFASTANDAWESOME)' 00:06:55.217 17:52:48 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@63 -- # nvme_name=nvme0n1 00:06:55.217 17:52:48 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@64 -- # sec_size_to_bytes nvme0n1 00:06:55.217 17:52:48 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- setup/common.sh@76 -- # local dev=nvme0n1 00:06:55.217 17:52:48 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- setup/common.sh@78 -- # [[ -e /sys/block/nvme0n1 ]] 00:06:55.217 17:52:48 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- setup/common.sh@80 -- # echo 536870912 00:06:55.217 17:52:48 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@64 -- # nvme_size=536870912 00:06:55.217 17:52:48 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@66 -- # mkdir -p /mnt/device 00:06:55.217 17:52:48 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@67 -- # (( nvme_size == malloc_size )) 00:06:55.217 17:52:48 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@68 -- # parted -s /dev/nvme0n1 mklabel gpt mkpart SPDK_TEST 0% 100% 00:06:55.217 17:52:48 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@69 -- # partprobe 00:06:55.217 17:52:48 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@70 -- # sleep 1 00:06:56.154 17:52:49 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@76 -- # '[' 4096 -eq 0 ']' 00:06:56.154 17:52:49 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@81 -- # run_test filesystem_in_capsule_ext4 nvmf_filesystem_create ext4 nvme0n1 00:06:56.154 17:52:49 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:06:56.154 17:52:49 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:56.154 17:52:49 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:56.154 ************************************ 00:06:56.154 START TEST filesystem_in_capsule_ext4 00:06:56.154 ************************************ 00:06:56.154 17:52:49 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@1123 -- # nvmf_filesystem_create ext4 nvme0n1 00:06:56.154 17:52:49 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@18 -- # fstype=ext4 00:06:56.154 17:52:49 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:06:56.154 17:52:49 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@21 -- # make_filesystem ext4 /dev/nvme0n1p1 00:06:56.154 17:52:49 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@924 -- # local fstype=ext4 00:06:56.154 17:52:49 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@925 -- # local dev_name=/dev/nvme0n1p1 00:06:56.154 17:52:49 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@926 -- # local i=0 00:06:56.154 17:52:49 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@927 -- # local force 00:06:56.154 17:52:49 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@929 -- # '[' ext4 = ext4 ']' 00:06:56.154 17:52:49 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@930 -- # force=-F 00:06:56.154 17:52:49 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@935 -- # mkfs.ext4 -F /dev/nvme0n1p1 00:06:56.154 mke2fs 1.46.5 (30-Dec-2021) 00:06:56.154 Discarding device blocks: 0/522240 done 00:06:56.154 Creating filesystem with 522240 1k blocks and 130560 inodes 00:06:56.154 Filesystem UUID: 65181e8f-6fb9-4c66-a4e3-092a3d65566f 00:06:56.154 Superblock backups stored on blocks: 00:06:56.154 8193, 24577, 40961, 57345, 73729, 204801, 221185, 401409 00:06:56.154 00:06:56.154 Allocating group tables: 0/64 done 00:06:56.154 Writing inode tables: 0/64 done 00:06:56.414 Creating journal (8192 blocks): done 00:06:57.348 Writing superblocks and filesystem accounting information: 0/64 done 00:06:57.348 00:06:57.348 17:52:50 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@943 -- # return 0 00:06:57.348 17:52:50 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@25 -- # sync 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@27 -- # sync 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@29 -- # i=0 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@30 -- # umount /mnt/device 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@37 -- # kill -0 437457 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:06:58.282 00:06:58.282 real 0m2.046s 00:06:58.282 user 0m0.030s 00:06:58.282 sys 0m0.060s 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_ext4 -- common/autotest_common.sh@10 -- # set +x 00:06:58.282 ************************************ 00:06:58.282 END TEST filesystem_in_capsule_ext4 00:06:58.282 ************************************ 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1142 -- # return 0 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@82 -- # run_test filesystem_in_capsule_btrfs nvmf_filesystem_create btrfs nvme0n1 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:58.282 ************************************ 00:06:58.282 START TEST filesystem_in_capsule_btrfs 00:06:58.282 ************************************ 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@1123 -- # nvmf_filesystem_create btrfs nvme0n1 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@18 -- # fstype=btrfs 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@21 -- # make_filesystem btrfs /dev/nvme0n1p1 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@924 -- # local fstype=btrfs 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@925 -- # local dev_name=/dev/nvme0n1p1 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@926 -- # local i=0 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@927 -- # local force 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@929 -- # '[' btrfs = ext4 ']' 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@932 -- # force=-f 00:06:58.282 17:52:51 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@935 -- # mkfs.btrfs -f /dev/nvme0n1p1 00:06:58.541 btrfs-progs v6.6.2 00:06:58.541 See https://btrfs.readthedocs.io for more information. 00:06:58.541 00:06:58.541 Performing full device TRIM /dev/nvme0n1p1 (510.00MiB) ... 00:06:58.541 NOTE: several default settings have changed in version 5.15, please make sure 00:06:58.541 this does not affect your deployments: 00:06:58.541 - DUP for metadata (-m dup) 00:06:58.541 - enabled no-holes (-O no-holes) 00:06:58.541 - enabled free-space-tree (-R free-space-tree) 00:06:58.541 00:06:58.541 Label: (null) 00:06:58.541 UUID: c054e141-37a1-49d3-a0e7-c50819442816 00:06:58.541 Node size: 16384 00:06:58.541 Sector size: 4096 00:06:58.541 Filesystem size: 510.00MiB 00:06:58.541 Block group profiles: 00:06:58.541 Data: single 8.00MiB 00:06:58.541 Metadata: DUP 32.00MiB 00:06:58.541 System: DUP 8.00MiB 00:06:58.541 SSD detected: yes 00:06:58.541 Zoned device: no 00:06:58.541 Incompat features: extref, skinny-metadata, no-holes, free-space-tree 00:06:58.541 Runtime features: free-space-tree 00:06:58.541 Checksum: crc32c 00:06:58.541 Number of devices: 1 00:06:58.541 Devices: 00:06:58.541 ID SIZE PATH 00:06:58.541 1 510.00MiB /dev/nvme0n1p1 00:06:58.541 00:06:58.541 17:52:52 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@943 -- # return 0 00:06:58.541 17:52:52 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@25 -- # sync 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@27 -- # sync 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@29 -- # i=0 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@30 -- # umount /mnt/device 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@37 -- # kill -0 437457 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:06:59.476 00:06:59.476 real 0m1.250s 00:06:59.476 user 0m0.016s 00:06:59.476 sys 0m0.135s 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@1124 -- # xtrace_disable 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_btrfs -- common/autotest_common.sh@10 -- # set +x 00:06:59.476 ************************************ 00:06:59.476 END TEST filesystem_in_capsule_btrfs 00:06:59.476 ************************************ 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1142 -- # return 0 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@83 -- # run_test filesystem_in_capsule_xfs nvmf_filesystem_create xfs nvme0n1 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1105 -- # xtrace_disable 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:06:59.476 ************************************ 00:06:59.476 START TEST filesystem_in_capsule_xfs 00:06:59.476 ************************************ 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@1123 -- # nvmf_filesystem_create xfs nvme0n1 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@18 -- # fstype=xfs 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@19 -- # nvme_name=nvme0n1 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@21 -- # make_filesystem xfs /dev/nvme0n1p1 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@924 -- # local fstype=xfs 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@925 -- # local dev_name=/dev/nvme0n1p1 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@926 -- # local i=0 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@927 -- # local force 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@929 -- # '[' xfs = ext4 ']' 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@932 -- # force=-f 00:06:59.476 17:52:53 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@935 -- # mkfs.xfs -f /dev/nvme0n1p1 00:06:59.734 meta-data=/dev/nvme0n1p1 isize=512 agcount=4, agsize=32640 blks 00:06:59.734 = sectsz=512 attr=2, projid32bit=1 00:06:59.734 = crc=1 finobt=1, sparse=1, rmapbt=0 00:06:59.734 = reflink=1 bigtime=1 inobtcount=1 nrext64=0 00:06:59.734 data = bsize=4096 blocks=130560, imaxpct=25 00:06:59.734 = sunit=0 swidth=0 blks 00:06:59.734 naming =version 2 bsize=4096 ascii-ci=0, ftype=1 00:06:59.734 log =internal log bsize=4096 blocks=16384, version=2 00:06:59.734 = sectsz=512 sunit=0 blks, lazy-count=1 00:06:59.734 realtime =none extsz=4096 blocks=0, rtextents=0 00:07:00.301 Discarding blocks...Done. 00:07:00.301 17:52:54 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@943 -- # return 0 00:07:00.301 17:52:54 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@23 -- # mount /dev/nvme0n1p1 /mnt/device 00:07:02.242 17:52:55 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@24 -- # touch /mnt/device/aaa 00:07:02.242 17:52:55 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@25 -- # sync 00:07:02.242 17:52:55 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@26 -- # rm /mnt/device/aaa 00:07:02.242 17:52:55 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@27 -- # sync 00:07:02.242 17:52:55 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@29 -- # i=0 00:07:02.242 17:52:55 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@30 -- # umount /mnt/device 00:07:02.242 17:52:55 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@37 -- # kill -0 437457 00:07:02.242 17:52:55 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@40 -- # lsblk -l -o NAME 00:07:02.242 17:52:55 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@40 -- # grep -q -w nvme0n1 00:07:02.242 17:52:55 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@43 -- # lsblk -l -o NAME 00:07:02.242 17:52:55 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- target/filesystem.sh@43 -- # grep -q -w nvme0n1p1 00:07:02.242 00:07:02.242 real 0m2.650s 00:07:02.242 user 0m0.023s 00:07:02.242 sys 0m0.072s 00:07:02.242 17:52:55 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@1124 -- # xtrace_disable 00:07:02.242 17:52:55 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule.filesystem_in_capsule_xfs -- common/autotest_common.sh@10 -- # set +x 00:07:02.242 ************************************ 00:07:02.242 END TEST filesystem_in_capsule_xfs 00:07:02.242 ************************************ 00:07:02.242 17:52:55 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1142 -- # return 0 00:07:02.242 17:52:55 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@91 -- # flock /dev/nvme0n1 parted -s /dev/nvme0n1 rm 1 00:07:02.501 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@93 -- # sync 00:07:02.501 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@94 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:07:02.760 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@95 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1219 -- # local i=0 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1231 -- # return 0 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@97 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@101 -- # killprocess 437457 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@948 -- # '[' -z 437457 ']' 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@952 -- # kill -0 437457 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@953 -- # uname 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 437457 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@966 -- # echo 'killing process with pid 437457' 00:07:02.760 killing process with pid 437457 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@967 -- # kill 437457 00:07:02.760 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@972 -- # wait 437457 00:07:03.019 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- target/filesystem.sh@102 -- # nvmfpid= 00:07:03.019 00:07:03.019 real 0m12.626s 00:07:03.019 user 0m49.602s 00:07:03.019 sys 0m1.248s 00:07:03.019 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@1124 -- # xtrace_disable 00:07:03.019 17:52:56 nvmf_tcp.nvmf_filesystem.nvmf_filesystem_in_capsule -- common/autotest_common.sh@10 -- # set +x 00:07:03.019 ************************************ 00:07:03.019 END TEST nvmf_filesystem_in_capsule 00:07:03.019 ************************************ 00:07:03.020 17:52:56 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@1142 -- # return 0 00:07:03.020 17:52:56 nvmf_tcp.nvmf_filesystem -- target/filesystem.sh@108 -- # nvmftestfini 00:07:03.020 17:52:56 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@488 -- # nvmfcleanup 00:07:03.020 17:52:56 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@117 -- # sync 00:07:03.020 17:52:56 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:07:03.020 17:52:56 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@120 -- # set +e 00:07:03.020 17:52:56 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@121 -- # for i in {1..20} 00:07:03.020 17:52:56 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:07:03.020 rmmod nvme_tcp 00:07:03.278 rmmod nvme_fabrics 00:07:03.278 rmmod nvme_keyring 00:07:03.278 17:52:56 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:07:03.278 17:52:56 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@124 -- # set -e 00:07:03.278 17:52:56 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@125 -- # return 0 00:07:03.278 17:52:56 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@489 -- # '[' -n '' ']' 00:07:03.278 17:52:56 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:07:03.278 17:52:56 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:07:03.278 17:52:56 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:07:03.278 17:52:56 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:07:03.278 17:52:56 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@278 -- # remove_spdk_ns 00:07:03.278 17:52:56 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:07:03.278 17:52:56 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:07:03.278 17:52:56 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:07:05.183 17:52:58 nvmf_tcp.nvmf_filesystem -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:07:05.183 00:07:05.183 real 0m34.486s 00:07:05.183 user 1m47.935s 00:07:05.183 sys 0m6.247s 00:07:05.183 17:52:58 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@1124 -- # xtrace_disable 00:07:05.183 17:52:58 nvmf_tcp.nvmf_filesystem -- common/autotest_common.sh@10 -- # set +x 00:07:05.183 ************************************ 00:07:05.183 END TEST nvmf_filesystem 00:07:05.183 ************************************ 00:07:05.183 17:52:58 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:07:05.183 17:52:58 nvmf_tcp -- nvmf/nvmf.sh@25 -- # run_test nvmf_target_discovery /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/discovery.sh --transport=tcp 00:07:05.183 17:52:58 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:07:05.183 17:52:58 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:07:05.183 17:52:58 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:07:05.442 ************************************ 00:07:05.442 START TEST nvmf_target_discovery 00:07:05.442 ************************************ 00:07:05.442 17:52:58 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/discovery.sh --transport=tcp 00:07:05.442 * Looking for test storage... 00:07:05.442 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@7 -- # uname -s 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- paths/export.sh@5 -- # export PATH 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@47 -- # : 0 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@51 -- # have_pci_nics=0 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@11 -- # NULL_BDEV_SIZE=102400 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@12 -- # NULL_BLOCK_SIZE=512 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@13 -- # NVMF_PORT_REFERRAL=4430 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@15 -- # hash nvme 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@20 -- # nvmftestinit 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@448 -- # prepare_net_devs 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@410 -- # local -g is_hw=no 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@412 -- # remove_spdk_ns 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@285 -- # xtrace_disable 00:07:05.442 17:52:59 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@291 -- # pci_devs=() 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@291 -- # local -a pci_devs 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@292 -- # pci_net_devs=() 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@293 -- # pci_drivers=() 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@293 -- # local -A pci_drivers 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@295 -- # net_devs=() 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@295 -- # local -ga net_devs 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@296 -- # e810=() 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@296 -- # local -ga e810 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@297 -- # x722=() 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@297 -- # local -ga x722 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@298 -- # mlx=() 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@298 -- # local -ga mlx 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:07:10.713 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:07:10.714 Found 0000:86:00.0 (0x8086 - 0x159b) 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:07:10.714 Found 0000:86:00.1 (0x8086 - 0x159b) 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@390 -- # [[ up == up ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:07:10.714 Found net devices under 0000:86:00.0: cvl_0_0 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@390 -- # [[ up == up ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:07:10.714 Found net devices under 0000:86:00.1: cvl_0_1 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@414 -- # is_hw=yes 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:07:10.714 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:07:10.974 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:07:10.974 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.170 ms 00:07:10.974 00:07:10.974 --- 10.0.0.2 ping statistics --- 00:07:10.974 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:07:10.974 rtt min/avg/max/mdev = 0.170/0.170/0.170/0.000 ms 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:07:10.974 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:07:10.974 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.233 ms 00:07:10.974 00:07:10.974 --- 10.0.0.1 ping statistics --- 00:07:10.974 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:07:10.974 rtt min/avg/max/mdev = 0.233/0.233/0.233/0.000 ms 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@422 -- # return 0 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@21 -- # nvmfappstart -m 0xF 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@722 -- # xtrace_disable 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@481 -- # nvmfpid=443268 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@482 -- # waitforlisten 443268 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@829 -- # '[' -z 443268 ']' 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@834 -- # local max_retries=100 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:10.974 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@838 -- # xtrace_disable 00:07:10.974 17:53:04 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:10.974 [2024-07-15 17:53:04.662821] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:07:10.974 [2024-07-15 17:53:04.662869] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:10.974 EAL: No free 2048 kB hugepages reported on node 1 00:07:11.233 [2024-07-15 17:53:04.723479] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:07:11.233 [2024-07-15 17:53:04.800129] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:07:11.233 [2024-07-15 17:53:04.800170] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:07:11.233 [2024-07-15 17:53:04.800176] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:07:11.233 [2024-07-15 17:53:04.800182] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:07:11.233 [2024-07-15 17:53:04.800187] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:07:11.233 [2024-07-15 17:53:04.800500] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:07:11.233 [2024-07-15 17:53:04.800581] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:07:11.233 [2024-07-15 17:53:04.800665] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:07:11.233 [2024-07-15 17:53:04.800666] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:07:11.803 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:07:11.803 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@862 -- # return 0 00:07:11.803 17:53:05 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:07:11.803 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@728 -- # xtrace_disable 00:07:11.803 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:11.803 17:53:05 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:07:11.803 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@23 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:07:11.803 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:11.803 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:11.803 [2024-07-15 17:53:05.515149] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:07:11.803 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:11.803 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@26 -- # seq 1 4 00:07:11.803 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:07:11.803 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null1 102400 512 00:07:11.803 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:11.803 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.062 Null1 00:07:12.062 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Null1 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.063 [2024-07-15 17:53:05.560718] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null2 102400 512 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.063 Null2 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK00000000000002 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Null2 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null3 102400 512 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.063 Null3 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode3 -a -s SPDK00000000000003 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode3 Null3 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode3 -t tcp -a 10.0.0.2 -s 4420 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@26 -- # for i in $(seq 1 4) 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@27 -- # rpc_cmd bdev_null_create Null4 102400 512 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.063 Null4 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode4 -a -s SPDK00000000000004 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode4 Null4 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@30 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode4 -t tcp -a 10.0.0.2 -s 4420 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@32 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@35 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 10.0.0.2 -s 4430 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.063 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@37 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -a 10.0.0.2 -s 4420 00:07:12.323 00:07:12.323 Discovery Log Number of Records 6, Generation counter 6 00:07:12.323 =====Discovery Log Entry 0====== 00:07:12.323 trtype: tcp 00:07:12.323 adrfam: ipv4 00:07:12.323 subtype: current discovery subsystem 00:07:12.323 treq: not required 00:07:12.323 portid: 0 00:07:12.323 trsvcid: 4420 00:07:12.323 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:07:12.323 traddr: 10.0.0.2 00:07:12.323 eflags: explicit discovery connections, duplicate discovery information 00:07:12.323 sectype: none 00:07:12.323 =====Discovery Log Entry 1====== 00:07:12.323 trtype: tcp 00:07:12.323 adrfam: ipv4 00:07:12.323 subtype: nvme subsystem 00:07:12.323 treq: not required 00:07:12.323 portid: 0 00:07:12.323 trsvcid: 4420 00:07:12.323 subnqn: nqn.2016-06.io.spdk:cnode1 00:07:12.323 traddr: 10.0.0.2 00:07:12.323 eflags: none 00:07:12.323 sectype: none 00:07:12.323 =====Discovery Log Entry 2====== 00:07:12.323 trtype: tcp 00:07:12.323 adrfam: ipv4 00:07:12.323 subtype: nvme subsystem 00:07:12.323 treq: not required 00:07:12.323 portid: 0 00:07:12.323 trsvcid: 4420 00:07:12.324 subnqn: nqn.2016-06.io.spdk:cnode2 00:07:12.324 traddr: 10.0.0.2 00:07:12.324 eflags: none 00:07:12.324 sectype: none 00:07:12.324 =====Discovery Log Entry 3====== 00:07:12.324 trtype: tcp 00:07:12.324 adrfam: ipv4 00:07:12.324 subtype: nvme subsystem 00:07:12.324 treq: not required 00:07:12.324 portid: 0 00:07:12.324 trsvcid: 4420 00:07:12.324 subnqn: nqn.2016-06.io.spdk:cnode3 00:07:12.324 traddr: 10.0.0.2 00:07:12.324 eflags: none 00:07:12.324 sectype: none 00:07:12.324 =====Discovery Log Entry 4====== 00:07:12.324 trtype: tcp 00:07:12.324 adrfam: ipv4 00:07:12.324 subtype: nvme subsystem 00:07:12.324 treq: not required 00:07:12.324 portid: 0 00:07:12.324 trsvcid: 4420 00:07:12.324 subnqn: nqn.2016-06.io.spdk:cnode4 00:07:12.324 traddr: 10.0.0.2 00:07:12.324 eflags: none 00:07:12.324 sectype: none 00:07:12.324 =====Discovery Log Entry 5====== 00:07:12.324 trtype: tcp 00:07:12.324 adrfam: ipv4 00:07:12.324 subtype: discovery subsystem referral 00:07:12.324 treq: not required 00:07:12.324 portid: 0 00:07:12.324 trsvcid: 4430 00:07:12.324 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:07:12.324 traddr: 10.0.0.2 00:07:12.324 eflags: none 00:07:12.324 sectype: none 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@39 -- # echo 'Perform nvmf subsystem discovery via RPC' 00:07:12.324 Perform nvmf subsystem discovery via RPC 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@40 -- # rpc_cmd nvmf_get_subsystems 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.324 [ 00:07:12.324 { 00:07:12.324 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:07:12.324 "subtype": "Discovery", 00:07:12.324 "listen_addresses": [ 00:07:12.324 { 00:07:12.324 "trtype": "TCP", 00:07:12.324 "adrfam": "IPv4", 00:07:12.324 "traddr": "10.0.0.2", 00:07:12.324 "trsvcid": "4420" 00:07:12.324 } 00:07:12.324 ], 00:07:12.324 "allow_any_host": true, 00:07:12.324 "hosts": [] 00:07:12.324 }, 00:07:12.324 { 00:07:12.324 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:07:12.324 "subtype": "NVMe", 00:07:12.324 "listen_addresses": [ 00:07:12.324 { 00:07:12.324 "trtype": "TCP", 00:07:12.324 "adrfam": "IPv4", 00:07:12.324 "traddr": "10.0.0.2", 00:07:12.324 "trsvcid": "4420" 00:07:12.324 } 00:07:12.324 ], 00:07:12.324 "allow_any_host": true, 00:07:12.324 "hosts": [], 00:07:12.324 "serial_number": "SPDK00000000000001", 00:07:12.324 "model_number": "SPDK bdev Controller", 00:07:12.324 "max_namespaces": 32, 00:07:12.324 "min_cntlid": 1, 00:07:12.324 "max_cntlid": 65519, 00:07:12.324 "namespaces": [ 00:07:12.324 { 00:07:12.324 "nsid": 1, 00:07:12.324 "bdev_name": "Null1", 00:07:12.324 "name": "Null1", 00:07:12.324 "nguid": "ED3DC991F51440ABA717D3FDFB8A2882", 00:07:12.324 "uuid": "ed3dc991-f514-40ab-a717-d3fdfb8a2882" 00:07:12.324 } 00:07:12.324 ] 00:07:12.324 }, 00:07:12.324 { 00:07:12.324 "nqn": "nqn.2016-06.io.spdk:cnode2", 00:07:12.324 "subtype": "NVMe", 00:07:12.324 "listen_addresses": [ 00:07:12.324 { 00:07:12.324 "trtype": "TCP", 00:07:12.324 "adrfam": "IPv4", 00:07:12.324 "traddr": "10.0.0.2", 00:07:12.324 "trsvcid": "4420" 00:07:12.324 } 00:07:12.324 ], 00:07:12.324 "allow_any_host": true, 00:07:12.324 "hosts": [], 00:07:12.324 "serial_number": "SPDK00000000000002", 00:07:12.324 "model_number": "SPDK bdev Controller", 00:07:12.324 "max_namespaces": 32, 00:07:12.324 "min_cntlid": 1, 00:07:12.324 "max_cntlid": 65519, 00:07:12.324 "namespaces": [ 00:07:12.324 { 00:07:12.324 "nsid": 1, 00:07:12.324 "bdev_name": "Null2", 00:07:12.324 "name": "Null2", 00:07:12.324 "nguid": "AD18B5944E734976B6C7160BDCDC9C15", 00:07:12.324 "uuid": "ad18b594-4e73-4976-b6c7-160bdcdc9c15" 00:07:12.324 } 00:07:12.324 ] 00:07:12.324 }, 00:07:12.324 { 00:07:12.324 "nqn": "nqn.2016-06.io.spdk:cnode3", 00:07:12.324 "subtype": "NVMe", 00:07:12.324 "listen_addresses": [ 00:07:12.324 { 00:07:12.324 "trtype": "TCP", 00:07:12.324 "adrfam": "IPv4", 00:07:12.324 "traddr": "10.0.0.2", 00:07:12.324 "trsvcid": "4420" 00:07:12.324 } 00:07:12.324 ], 00:07:12.324 "allow_any_host": true, 00:07:12.324 "hosts": [], 00:07:12.324 "serial_number": "SPDK00000000000003", 00:07:12.324 "model_number": "SPDK bdev Controller", 00:07:12.324 "max_namespaces": 32, 00:07:12.324 "min_cntlid": 1, 00:07:12.324 "max_cntlid": 65519, 00:07:12.324 "namespaces": [ 00:07:12.324 { 00:07:12.324 "nsid": 1, 00:07:12.324 "bdev_name": "Null3", 00:07:12.324 "name": "Null3", 00:07:12.324 "nguid": "9DB139086BB54C438346FB22849ECD98", 00:07:12.324 "uuid": "9db13908-6bb5-4c43-8346-fb22849ecd98" 00:07:12.324 } 00:07:12.324 ] 00:07:12.324 }, 00:07:12.324 { 00:07:12.324 "nqn": "nqn.2016-06.io.spdk:cnode4", 00:07:12.324 "subtype": "NVMe", 00:07:12.324 "listen_addresses": [ 00:07:12.324 { 00:07:12.324 "trtype": "TCP", 00:07:12.324 "adrfam": "IPv4", 00:07:12.324 "traddr": "10.0.0.2", 00:07:12.324 "trsvcid": "4420" 00:07:12.324 } 00:07:12.324 ], 00:07:12.324 "allow_any_host": true, 00:07:12.324 "hosts": [], 00:07:12.324 "serial_number": "SPDK00000000000004", 00:07:12.324 "model_number": "SPDK bdev Controller", 00:07:12.324 "max_namespaces": 32, 00:07:12.324 "min_cntlid": 1, 00:07:12.324 "max_cntlid": 65519, 00:07:12.324 "namespaces": [ 00:07:12.324 { 00:07:12.324 "nsid": 1, 00:07:12.324 "bdev_name": "Null4", 00:07:12.324 "name": "Null4", 00:07:12.324 "nguid": "5F3AAA01DFB443EE98A549237ECF93E0", 00:07:12.324 "uuid": "5f3aaa01-dfb4-43ee-98a5-49237ecf93e0" 00:07:12.324 } 00:07:12.324 ] 00:07:12.324 } 00:07:12.324 ] 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@42 -- # seq 1 4 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null1 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode2 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null2 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode3 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null3 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@42 -- # for i in $(seq 1 4) 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@43 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode4 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@44 -- # rpc_cmd bdev_null_delete Null4 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@47 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 10.0.0.2 -s 4430 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@49 -- # rpc_cmd bdev_get_bdevs 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@49 -- # jq -r '.[].name' 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@49 -- # check_bdevs= 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@50 -- # '[' -n '' ']' 00:07:12.324 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@55 -- # trap - SIGINT SIGTERM EXIT 00:07:12.325 17:53:05 nvmf_tcp.nvmf_target_discovery -- target/discovery.sh@57 -- # nvmftestfini 00:07:12.325 17:53:05 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@488 -- # nvmfcleanup 00:07:12.325 17:53:05 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@117 -- # sync 00:07:12.325 17:53:05 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:07:12.325 17:53:05 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@120 -- # set +e 00:07:12.325 17:53:05 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@121 -- # for i in {1..20} 00:07:12.325 17:53:05 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:07:12.325 rmmod nvme_tcp 00:07:12.325 rmmod nvme_fabrics 00:07:12.325 rmmod nvme_keyring 00:07:12.325 17:53:06 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:07:12.325 17:53:06 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@124 -- # set -e 00:07:12.325 17:53:06 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@125 -- # return 0 00:07:12.325 17:53:06 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@489 -- # '[' -n 443268 ']' 00:07:12.325 17:53:06 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@490 -- # killprocess 443268 00:07:12.325 17:53:06 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@948 -- # '[' -z 443268 ']' 00:07:12.325 17:53:06 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@952 -- # kill -0 443268 00:07:12.325 17:53:06 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@953 -- # uname 00:07:12.325 17:53:06 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:07:12.325 17:53:06 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 443268 00:07:12.584 17:53:06 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:07:12.584 17:53:06 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:07:12.584 17:53:06 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@966 -- # echo 'killing process with pid 443268' 00:07:12.584 killing process with pid 443268 00:07:12.584 17:53:06 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@967 -- # kill 443268 00:07:12.584 17:53:06 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@972 -- # wait 443268 00:07:12.584 17:53:06 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:07:12.584 17:53:06 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:07:12.584 17:53:06 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:07:12.584 17:53:06 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:07:12.584 17:53:06 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@278 -- # remove_spdk_ns 00:07:12.584 17:53:06 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:07:12.584 17:53:06 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:07:12.584 17:53:06 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:07:15.122 17:53:08 nvmf_tcp.nvmf_target_discovery -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:07:15.122 00:07:15.122 real 0m9.397s 00:07:15.122 user 0m7.581s 00:07:15.122 sys 0m4.531s 00:07:15.122 17:53:08 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@1124 -- # xtrace_disable 00:07:15.122 17:53:08 nvmf_tcp.nvmf_target_discovery -- common/autotest_common.sh@10 -- # set +x 00:07:15.122 ************************************ 00:07:15.122 END TEST nvmf_target_discovery 00:07:15.122 ************************************ 00:07:15.122 17:53:08 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:07:15.122 17:53:08 nvmf_tcp -- nvmf/nvmf.sh@26 -- # run_test nvmf_referrals /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/referrals.sh --transport=tcp 00:07:15.122 17:53:08 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:07:15.122 17:53:08 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:07:15.122 17:53:08 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:07:15.122 ************************************ 00:07:15.122 START TEST nvmf_referrals 00:07:15.122 ************************************ 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/referrals.sh --transport=tcp 00:07:15.122 * Looking for test storage... 00:07:15.122 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- target/referrals.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@7 -- # uname -s 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- paths/export.sh@5 -- # export PATH 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@47 -- # : 0 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@51 -- # have_pci_nics=0 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- target/referrals.sh@11 -- # NVMF_REFERRAL_IP_1=127.0.0.2 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- target/referrals.sh@12 -- # NVMF_REFERRAL_IP_2=127.0.0.3 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- target/referrals.sh@13 -- # NVMF_REFERRAL_IP_3=127.0.0.4 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- target/referrals.sh@14 -- # NVMF_PORT_REFERRAL=4430 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- target/referrals.sh@15 -- # DISCOVERY_NQN=nqn.2014-08.org.nvmexpress.discovery 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- target/referrals.sh@16 -- # NQN=nqn.2016-06.io.spdk:cnode1 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- target/referrals.sh@37 -- # nvmftestinit 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@448 -- # prepare_net_devs 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@410 -- # local -g is_hw=no 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@412 -- # remove_spdk_ns 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@285 -- # xtrace_disable 00:07:15.122 17:53:08 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@291 -- # pci_devs=() 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@291 -- # local -a pci_devs 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@292 -- # pci_net_devs=() 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@293 -- # pci_drivers=() 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@293 -- # local -A pci_drivers 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@295 -- # net_devs=() 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@295 -- # local -ga net_devs 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@296 -- # e810=() 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@296 -- # local -ga e810 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@297 -- # x722=() 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@297 -- # local -ga x722 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@298 -- # mlx=() 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@298 -- # local -ga mlx 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:07:20.399 Found 0000:86:00.0 (0x8086 - 0x159b) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:07:20.399 Found 0000:86:00.1 (0x8086 - 0x159b) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@390 -- # [[ up == up ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:07:20.399 Found net devices under 0000:86:00.0: cvl_0_0 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@390 -- # [[ up == up ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:07:20.399 Found net devices under 0000:86:00.1: cvl_0_1 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@414 -- # is_hw=yes 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:07:20.399 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:07:20.399 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:07:20.399 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.194 ms 00:07:20.399 00:07:20.399 --- 10.0.0.2 ping statistics --- 00:07:20.400 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:07:20.400 rtt min/avg/max/mdev = 0.194/0.194/0.194/0.000 ms 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:07:20.400 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:07:20.400 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.222 ms 00:07:20.400 00:07:20.400 --- 10.0.0.1 ping statistics --- 00:07:20.400 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:07:20.400 rtt min/avg/max/mdev = 0.222/0.222/0.222/0.000 ms 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@422 -- # return 0 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- target/referrals.sh@38 -- # nvmfappstart -m 0xF 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@722 -- # xtrace_disable 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@481 -- # nvmfpid=447043 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@482 -- # waitforlisten 447043 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@829 -- # '[' -z 447043 ']' 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@834 -- # local max_retries=100 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:20.400 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@838 -- # xtrace_disable 00:07:20.400 17:53:13 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:20.400 [2024-07-15 17:53:13.972756] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:07:20.400 [2024-07-15 17:53:13.972800] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:20.400 EAL: No free 2048 kB hugepages reported on node 1 00:07:20.400 [2024-07-15 17:53:14.033362] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:07:20.400 [2024-07-15 17:53:14.114214] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:07:20.400 [2024-07-15 17:53:14.114253] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:07:20.400 [2024-07-15 17:53:14.114261] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:07:20.400 [2024-07-15 17:53:14.114267] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:07:20.400 [2024-07-15 17:53:14.114272] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:07:20.400 [2024-07-15 17:53:14.114487] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:07:20.400 [2024-07-15 17:53:14.114566] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:07:20.400 [2024-07-15 17:53:14.114686] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:07:20.400 [2024-07-15 17:53:14.114688] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@862 -- # return 0 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@728 -- # xtrace_disable 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@40 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:21.335 [2024-07-15 17:53:14.818068] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@41 -- # rpc_cmd nvmf_subsystem_add_listener -t tcp -a 10.0.0.2 -s 8009 discovery 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:21.335 [2024-07-15 17:53:14.831411] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 8009 *** 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@44 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.2 -s 4430 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@45 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.3 -s 4430 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@46 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.4 -s 4430 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@48 -- # jq length 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@48 -- # rpc_cmd nvmf_discovery_get_referrals 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@48 -- # (( 3 == 3 )) 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@49 -- # get_referral_ips rpc 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@19 -- # [[ rpc == \r\p\c ]] 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@21 -- # rpc_cmd nvmf_discovery_get_referrals 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@21 -- # jq -r '.[].address.traddr' 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@21 -- # sort 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@21 -- # echo 127.0.0.2 127.0.0.3 127.0.0.4 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@49 -- # [[ 127.0.0.2 127.0.0.3 127.0.0.4 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\3\ \1\2\7\.\0\.\0\.\4 ]] 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@50 -- # get_referral_ips nvme 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -a 10.0.0.2 -s 8009 -o json 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:07:21.335 17:53:14 nvmf_tcp.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@26 -- # echo 127.0.0.2 127.0.0.3 127.0.0.4 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@50 -- # [[ 127.0.0.2 127.0.0.3 127.0.0.4 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\3\ \1\2\7\.\0\.\0\.\4 ]] 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@52 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.2 -s 4430 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@53 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.3 -s 4430 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@54 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.4 -s 4430 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@56 -- # rpc_cmd nvmf_discovery_get_referrals 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@56 -- # jq length 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@56 -- # (( 0 == 0 )) 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@57 -- # get_referral_ips nvme 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -a 10.0.0.2 -s 8009 -o json 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@26 -- # echo 00:07:21.594 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@57 -- # [[ '' == '' ]] 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@60 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.2 -s 4430 -n discovery 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@62 -- # rpc_cmd nvmf_discovery_add_referral -t tcp -a 127.0.0.2 -s 4430 -n nqn.2016-06.io.spdk:cnode1 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@65 -- # get_referral_ips rpc 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@19 -- # [[ rpc == \r\p\c ]] 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@21 -- # rpc_cmd nvmf_discovery_get_referrals 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@21 -- # jq -r '.[].address.traddr' 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@21 -- # sort 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@21 -- # echo 127.0.0.2 127.0.0.2 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@65 -- # [[ 127.0.0.2 127.0.0.2 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\2 ]] 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@66 -- # get_referral_ips nvme 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -a 10.0.0.2 -s 8009 -o json 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@26 -- # echo 127.0.0.2 127.0.0.2 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@66 -- # [[ 127.0.0.2 127.0.0.2 == \1\2\7\.\0\.\0\.\2\ \1\2\7\.\0\.\0\.\2 ]] 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@67 -- # get_discovery_entries 'nvme subsystem' 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@67 -- # jq -r .subnqn 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@31 -- # local 'subtype=nvme subsystem' 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -a 10.0.0.2 -s 8009 -o json 00:07:21.854 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "nvme subsystem")' 00:07:22.113 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@67 -- # [[ nqn.2016-06.io.spdk:cnode1 == \n\q\n\.\2\0\1\6\-\0\6\.\i\o\.\s\p\d\k\:\c\n\o\d\e\1 ]] 00:07:22.113 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@68 -- # get_discovery_entries 'discovery subsystem referral' 00:07:22.113 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@68 -- # jq -r .subnqn 00:07:22.113 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@31 -- # local 'subtype=discovery subsystem referral' 00:07:22.113 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -a 10.0.0.2 -s 8009 -o json 00:07:22.113 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "discovery subsystem referral")' 00:07:22.113 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@68 -- # [[ nqn.2014-08.org.nvmexpress.discovery == \n\q\n\.\2\0\1\4\-\0\8\.\o\r\g\.\n\v\m\e\x\p\r\e\s\s\.\d\i\s\c\o\v\e\r\y ]] 00:07:22.113 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@71 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.2 -s 4430 -n nqn.2016-06.io.spdk:cnode1 00:07:22.113 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:22.113 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:22.113 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:22.113 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@73 -- # get_referral_ips rpc 00:07:22.113 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@19 -- # [[ rpc == \r\p\c ]] 00:07:22.113 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@21 -- # rpc_cmd nvmf_discovery_get_referrals 00:07:22.113 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:22.113 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@21 -- # jq -r '.[].address.traddr' 00:07:22.113 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:22.113 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@21 -- # sort 00:07:22.113 17:53:15 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:22.372 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@21 -- # echo 127.0.0.2 00:07:22.372 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@73 -- # [[ 127.0.0.2 == \1\2\7\.\0\.\0\.\2 ]] 00:07:22.372 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@74 -- # get_referral_ips nvme 00:07:22.372 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:07:22.372 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:07:22.372 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:07:22.372 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -a 10.0.0.2 -s 8009 -o json 00:07:22.372 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:07:22.372 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@26 -- # echo 127.0.0.2 00:07:22.372 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@74 -- # [[ 127.0.0.2 == \1\2\7\.\0\.\0\.\2 ]] 00:07:22.372 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@75 -- # get_discovery_entries 'nvme subsystem' 00:07:22.372 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@75 -- # jq -r .subnqn 00:07:22.372 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@31 -- # local 'subtype=nvme subsystem' 00:07:22.372 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -a 10.0.0.2 -s 8009 -o json 00:07:22.372 17:53:15 nvmf_tcp.nvmf_referrals -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "nvme subsystem")' 00:07:22.372 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@75 -- # [[ '' == '' ]] 00:07:22.372 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@76 -- # jq -r .subnqn 00:07:22.372 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@76 -- # get_discovery_entries 'discovery subsystem referral' 00:07:22.372 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@31 -- # local 'subtype=discovery subsystem referral' 00:07:22.372 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@33 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -a 10.0.0.2 -s 8009 -o json 00:07:22.372 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@34 -- # jq '.records[] | select(.subtype == "discovery subsystem referral")' 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@76 -- # [[ nqn.2014-08.org.nvmexpress.discovery == \n\q\n\.\2\0\1\4\-\0\8\.\o\r\g\.\n\v\m\e\x\p\r\e\s\s\.\d\i\s\c\o\v\e\r\y ]] 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@79 -- # rpc_cmd nvmf_discovery_remove_referral -t tcp -a 127.0.0.2 -s 4430 -n nqn.2014-08.org.nvmexpress.discovery 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@82 -- # rpc_cmd nvmf_discovery_get_referrals 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@82 -- # jq length 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@82 -- # (( 0 == 0 )) 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@83 -- # get_referral_ips nvme 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@19 -- # [[ nvme == \r\p\c ]] 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@22 -- # [[ nvme == \n\v\m\e ]] 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@26 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -a 10.0.0.2 -s 8009 -o json 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@26 -- # jq -r '.records[] | select(.subtype != "current discovery subsystem").traddr' 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@26 -- # sort 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@26 -- # echo 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@83 -- # [[ '' == '' ]] 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@85 -- # trap - SIGINT SIGTERM EXIT 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- target/referrals.sh@86 -- # nvmftestfini 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@488 -- # nvmfcleanup 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@117 -- # sync 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@120 -- # set +e 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@121 -- # for i in {1..20} 00:07:22.632 17:53:16 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:07:22.632 rmmod nvme_tcp 00:07:22.632 rmmod nvme_fabrics 00:07:22.892 rmmod nvme_keyring 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@124 -- # set -e 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@125 -- # return 0 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@489 -- # '[' -n 447043 ']' 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@490 -- # killprocess 447043 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@948 -- # '[' -z 447043 ']' 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@952 -- # kill -0 447043 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@953 -- # uname 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 447043 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@966 -- # echo 'killing process with pid 447043' 00:07:22.892 killing process with pid 447043 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@967 -- # kill 447043 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@972 -- # wait 447043 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@278 -- # remove_spdk_ns 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:07:22.892 17:53:16 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:07:25.474 17:53:18 nvmf_tcp.nvmf_referrals -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:07:25.474 00:07:25.474 real 0m10.268s 00:07:25.474 user 0m12.340s 00:07:25.474 sys 0m4.693s 00:07:25.474 17:53:18 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@1124 -- # xtrace_disable 00:07:25.474 17:53:18 nvmf_tcp.nvmf_referrals -- common/autotest_common.sh@10 -- # set +x 00:07:25.474 ************************************ 00:07:25.474 END TEST nvmf_referrals 00:07:25.474 ************************************ 00:07:25.474 17:53:18 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:07:25.474 17:53:18 nvmf_tcp -- nvmf/nvmf.sh@27 -- # run_test nvmf_connect_disconnect /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_disconnect.sh --transport=tcp 00:07:25.474 17:53:18 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:07:25.474 17:53:18 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:07:25.474 17:53:18 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:07:25.474 ************************************ 00:07:25.474 START TEST nvmf_connect_disconnect 00:07:25.474 ************************************ 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_disconnect.sh --transport=tcp 00:07:25.474 * Looking for test storage... 00:07:25.474 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- target/connect_disconnect.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@7 -- # uname -s 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:25.474 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- paths/export.sh@5 -- # export PATH 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@47 -- # : 0 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@51 -- # have_pci_nics=0 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- target/connect_disconnect.sh@11 -- # MALLOC_BDEV_SIZE=64 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- target/connect_disconnect.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- target/connect_disconnect.sh@15 -- # nvmftestinit 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@448 -- # prepare_net_devs 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@410 -- # local -g is_hw=no 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@412 -- # remove_spdk_ns 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@285 -- # xtrace_disable 00:07:25.475 17:53:18 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@291 -- # pci_devs=() 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@291 -- # local -a pci_devs 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@292 -- # pci_net_devs=() 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@293 -- # pci_drivers=() 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@293 -- # local -A pci_drivers 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@295 -- # net_devs=() 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@295 -- # local -ga net_devs 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@296 -- # e810=() 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@296 -- # local -ga e810 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@297 -- # x722=() 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@297 -- # local -ga x722 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@298 -- # mlx=() 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@298 -- # local -ga mlx 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:07:30.793 Found 0000:86:00.0 (0x8086 - 0x159b) 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:07:30.793 Found 0000:86:00.1 (0x8086 - 0x159b) 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@390 -- # [[ up == up ]] 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:07:30.793 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:07:30.793 Found net devices under 0000:86:00.0: cvl_0_0 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@390 -- # [[ up == up ]] 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:07:30.794 Found net devices under 0000:86:00.1: cvl_0_1 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@414 -- # is_hw=yes 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:07:30.794 17:53:23 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:07:30.794 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:07:30.794 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.276 ms 00:07:30.794 00:07:30.794 --- 10.0.0.2 ping statistics --- 00:07:30.794 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:07:30.794 rtt min/avg/max/mdev = 0.276/0.276/0.276/0.000 ms 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:07:30.794 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:07:30.794 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.097 ms 00:07:30.794 00:07:30.794 --- 10.0.0.1 ping statistics --- 00:07:30.794 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:07:30.794 rtt min/avg/max/mdev = 0.097/0.097/0.097/0.000 ms 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@422 -- # return 0 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- target/connect_disconnect.sh@16 -- # nvmfappstart -m 0xF 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@722 -- # xtrace_disable 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@481 -- # nvmfpid=450902 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@482 -- # waitforlisten 450902 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@829 -- # '[' -z 450902 ']' 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@834 -- # local max_retries=100 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:30.794 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@838 -- # xtrace_disable 00:07:30.794 17:53:24 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:07:30.794 [2024-07-15 17:53:24.274954] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:07:30.794 [2024-07-15 17:53:24.274994] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:30.794 EAL: No free 2048 kB hugepages reported on node 1 00:07:30.794 [2024-07-15 17:53:24.332416] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:07:30.794 [2024-07-15 17:53:24.405070] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:07:30.794 [2024-07-15 17:53:24.405113] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:07:30.794 [2024-07-15 17:53:24.405120] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:07:30.794 [2024-07-15 17:53:24.405126] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:07:30.794 [2024-07-15 17:53:24.405131] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:07:30.794 [2024-07-15 17:53:24.405183] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:07:30.794 [2024-07-15 17:53:24.405283] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:07:30.794 [2024-07-15 17:53:24.405305] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:07:30.794 [2024-07-15 17:53:24.405307] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:07:31.361 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:07:31.361 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@862 -- # return 0 00:07:31.361 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:07:31.361 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@728 -- # xtrace_disable 00:07:31.361 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- target/connect_disconnect.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -c 0 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:07:31.620 [2024-07-15 17:53:25.126340] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- target/connect_disconnect.sh@20 -- # rpc_cmd bdev_malloc_create 64 512 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- target/connect_disconnect.sh@20 -- # bdev=Malloc0 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- target/connect_disconnect.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- target/connect_disconnect.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- target/connect_disconnect.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@559 -- # xtrace_disable 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:07:31.620 [2024-07-15 17:53:25.178247] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- target/connect_disconnect.sh@26 -- # '[' 0 -eq 1 ']' 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- target/connect_disconnect.sh@31 -- # num_iterations=5 00:07:31.620 17:53:25 nvmf_tcp.nvmf_connect_disconnect -- target/connect_disconnect.sh@34 -- # set +x 00:07:34.908 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:07:38.196 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:07:41.485 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:07:44.773 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:07:48.062 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- target/connect_disconnect.sh@43 -- # trap - SIGINT SIGTERM EXIT 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- target/connect_disconnect.sh@45 -- # nvmftestfini 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@488 -- # nvmfcleanup 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@117 -- # sync 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@120 -- # set +e 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@121 -- # for i in {1..20} 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:07:48.062 rmmod nvme_tcp 00:07:48.062 rmmod nvme_fabrics 00:07:48.062 rmmod nvme_keyring 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@124 -- # set -e 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@125 -- # return 0 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@489 -- # '[' -n 450902 ']' 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@490 -- # killprocess 450902 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@948 -- # '[' -z 450902 ']' 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@952 -- # kill -0 450902 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@953 -- # uname 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 450902 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@966 -- # echo 'killing process with pid 450902' 00:07:48.062 killing process with pid 450902 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@967 -- # kill 450902 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@972 -- # wait 450902 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@278 -- # remove_spdk_ns 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:07:48.062 17:53:41 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:07:50.050 17:53:43 nvmf_tcp.nvmf_connect_disconnect -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:07:50.050 00:07:50.050 real 0m25.009s 00:07:50.050 user 1m10.238s 00:07:50.050 sys 0m5.203s 00:07:50.050 17:53:43 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@1124 -- # xtrace_disable 00:07:50.050 17:53:43 nvmf_tcp.nvmf_connect_disconnect -- common/autotest_common.sh@10 -- # set +x 00:07:50.050 ************************************ 00:07:50.050 END TEST nvmf_connect_disconnect 00:07:50.050 ************************************ 00:07:50.310 17:53:43 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:07:50.310 17:53:43 nvmf_tcp -- nvmf/nvmf.sh@28 -- # run_test nvmf_multitarget /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget.sh --transport=tcp 00:07:50.310 17:53:43 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:07:50.310 17:53:43 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:07:50.310 17:53:43 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:07:50.310 ************************************ 00:07:50.310 START TEST nvmf_multitarget 00:07:50.310 ************************************ 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget.sh --transport=tcp 00:07:50.310 * Looking for test storage... 00:07:50.310 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- target/multitarget.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@7 -- # uname -s 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- paths/export.sh@5 -- # export PATH 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@47 -- # : 0 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@51 -- # have_pci_nics=0 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- target/multitarget.sh@13 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- target/multitarget.sh@15 -- # nvmftestinit 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@448 -- # prepare_net_devs 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@410 -- # local -g is_hw=no 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@412 -- # remove_spdk_ns 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@285 -- # xtrace_disable 00:07:50.310 17:53:43 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:07:55.589 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:07:55.589 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@291 -- # pci_devs=() 00:07:55.589 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@291 -- # local -a pci_devs 00:07:55.589 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@292 -- # pci_net_devs=() 00:07:55.589 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:07:55.589 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@293 -- # pci_drivers=() 00:07:55.589 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@293 -- # local -A pci_drivers 00:07:55.589 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@295 -- # net_devs=() 00:07:55.589 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@295 -- # local -ga net_devs 00:07:55.589 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@296 -- # e810=() 00:07:55.589 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@296 -- # local -ga e810 00:07:55.589 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@297 -- # x722=() 00:07:55.589 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@297 -- # local -ga x722 00:07:55.589 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@298 -- # mlx=() 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@298 -- # local -ga mlx 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:07:55.590 Found 0000:86:00.0 (0x8086 - 0x159b) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:07:55.590 Found 0000:86:00.1 (0x8086 - 0x159b) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@390 -- # [[ up == up ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:07:55.590 Found net devices under 0000:86:00.0: cvl_0_0 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@390 -- # [[ up == up ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:07:55.590 Found net devices under 0000:86:00.1: cvl_0_1 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@414 -- # is_hw=yes 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:07:55.590 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:07:55.590 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.170 ms 00:07:55.590 00:07:55.590 --- 10.0.0.2 ping statistics --- 00:07:55.590 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:07:55.590 rtt min/avg/max/mdev = 0.170/0.170/0.170/0.000 ms 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:07:55.590 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:07:55.590 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.220 ms 00:07:55.590 00:07:55.590 --- 10.0.0.1 ping statistics --- 00:07:55.590 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:07:55.590 rtt min/avg/max/mdev = 0.220/0.220/0.220/0.000 ms 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@422 -- # return 0 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- target/multitarget.sh@16 -- # nvmfappstart -m 0xF 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@722 -- # xtrace_disable 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@481 -- # nvmfpid=457293 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@482 -- # waitforlisten 457293 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@829 -- # '[' -z 457293 ']' 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@834 -- # local max_retries=100 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:07:55.590 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@838 -- # xtrace_disable 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:07:55.590 17:53:48 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:07:55.590 [2024-07-15 17:53:48.987509] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:07:55.590 [2024-07-15 17:53:48.987555] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:07:55.590 EAL: No free 2048 kB hugepages reported on node 1 00:07:55.590 [2024-07-15 17:53:49.048565] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:07:55.590 [2024-07-15 17:53:49.132500] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:07:55.590 [2024-07-15 17:53:49.132533] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:07:55.590 [2024-07-15 17:53:49.132540] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:07:55.590 [2024-07-15 17:53:49.132547] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:07:55.590 [2024-07-15 17:53:49.132552] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:07:55.590 [2024-07-15 17:53:49.132593] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:07:55.590 [2024-07-15 17:53:49.132609] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:07:55.590 [2024-07-15 17:53:49.132704] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:07:55.590 [2024-07-15 17:53:49.132706] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:07:56.160 17:53:49 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:07:56.160 17:53:49 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@862 -- # return 0 00:07:56.160 17:53:49 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:07:56.160 17:53:49 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@728 -- # xtrace_disable 00:07:56.160 17:53:49 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:07:56.160 17:53:49 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:07:56.160 17:53:49 nvmf_tcp.nvmf_multitarget -- target/multitarget.sh@18 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini $1; exit 1' SIGINT SIGTERM EXIT 00:07:56.160 17:53:49 nvmf_tcp.nvmf_multitarget -- target/multitarget.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_get_targets 00:07:56.160 17:53:49 nvmf_tcp.nvmf_multitarget -- target/multitarget.sh@21 -- # jq length 00:07:56.420 17:53:49 nvmf_tcp.nvmf_multitarget -- target/multitarget.sh@21 -- # '[' 1 '!=' 1 ']' 00:07:56.420 17:53:49 nvmf_tcp.nvmf_multitarget -- target/multitarget.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_create_target -n nvmf_tgt_1 -s 32 00:07:56.420 "nvmf_tgt_1" 00:07:56.420 17:53:50 nvmf_tcp.nvmf_multitarget -- target/multitarget.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_create_target -n nvmf_tgt_2 -s 32 00:07:56.420 "nvmf_tgt_2" 00:07:56.679 17:53:50 nvmf_tcp.nvmf_multitarget -- target/multitarget.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_get_targets 00:07:56.679 17:53:50 nvmf_tcp.nvmf_multitarget -- target/multitarget.sh@28 -- # jq length 00:07:56.679 17:53:50 nvmf_tcp.nvmf_multitarget -- target/multitarget.sh@28 -- # '[' 3 '!=' 3 ']' 00:07:56.679 17:53:50 nvmf_tcp.nvmf_multitarget -- target/multitarget.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_delete_target -n nvmf_tgt_1 00:07:56.679 true 00:07:56.679 17:53:50 nvmf_tcp.nvmf_multitarget -- target/multitarget.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_delete_target -n nvmf_tgt_2 00:07:56.939 true 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- target/multitarget.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py nvmf_get_targets 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- target/multitarget.sh@35 -- # jq length 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- target/multitarget.sh@35 -- # '[' 1 '!=' 1 ']' 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- target/multitarget.sh@39 -- # trap - SIGINT SIGTERM EXIT 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- target/multitarget.sh@41 -- # nvmftestfini 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@488 -- # nvmfcleanup 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@117 -- # sync 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@120 -- # set +e 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@121 -- # for i in {1..20} 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:07:56.939 rmmod nvme_tcp 00:07:56.939 rmmod nvme_fabrics 00:07:56.939 rmmod nvme_keyring 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@124 -- # set -e 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@125 -- # return 0 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@489 -- # '[' -n 457293 ']' 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@490 -- # killprocess 457293 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@948 -- # '[' -z 457293 ']' 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@952 -- # kill -0 457293 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@953 -- # uname 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:07:56.939 17:53:50 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 457293 00:07:57.198 17:53:50 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:07:57.198 17:53:50 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:07:57.198 17:53:50 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@966 -- # echo 'killing process with pid 457293' 00:07:57.198 killing process with pid 457293 00:07:57.198 17:53:50 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@967 -- # kill 457293 00:07:57.198 17:53:50 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@972 -- # wait 457293 00:07:57.198 17:53:50 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:07:57.198 17:53:50 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:07:57.198 17:53:50 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:07:57.198 17:53:50 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:07:57.198 17:53:50 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@278 -- # remove_spdk_ns 00:07:57.198 17:53:50 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:07:57.198 17:53:50 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:07:57.198 17:53:50 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:07:59.736 17:53:52 nvmf_tcp.nvmf_multitarget -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:07:59.736 00:07:59.736 real 0m9.100s 00:07:59.736 user 0m8.924s 00:07:59.736 sys 0m4.167s 00:07:59.736 17:53:52 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@1124 -- # xtrace_disable 00:07:59.736 17:53:52 nvmf_tcp.nvmf_multitarget -- common/autotest_common.sh@10 -- # set +x 00:07:59.736 ************************************ 00:07:59.736 END TEST nvmf_multitarget 00:07:59.736 ************************************ 00:07:59.736 17:53:52 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:07:59.736 17:53:52 nvmf_tcp -- nvmf/nvmf.sh@29 -- # run_test nvmf_rpc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.sh --transport=tcp 00:07:59.736 17:53:52 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:07:59.736 17:53:52 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:07:59.736 17:53:52 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:07:59.736 ************************************ 00:07:59.736 START TEST nvmf_rpc 00:07:59.736 ************************************ 00:07:59.736 17:53:52 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.sh --transport=tcp 00:07:59.736 * Looking for test storage... 00:07:59.736 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- target/rpc.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@7 -- # uname -s 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- paths/export.sh@5 -- # export PATH 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@47 -- # : 0 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@51 -- # have_pci_nics=0 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- target/rpc.sh@11 -- # loops=5 00:07:59.736 17:53:53 nvmf_tcp.nvmf_rpc -- target/rpc.sh@23 -- # nvmftestinit 00:07:59.737 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:07:59.737 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:07:59.737 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@448 -- # prepare_net_devs 00:07:59.737 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@410 -- # local -g is_hw=no 00:07:59.737 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@412 -- # remove_spdk_ns 00:07:59.737 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:07:59.737 17:53:53 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:07:59.737 17:53:53 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:07:59.737 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:07:59.737 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:07:59.737 17:53:53 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@285 -- # xtrace_disable 00:07:59.737 17:53:53 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@291 -- # pci_devs=() 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@291 -- # local -a pci_devs 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@292 -- # pci_net_devs=() 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@293 -- # pci_drivers=() 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@293 -- # local -A pci_drivers 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@295 -- # net_devs=() 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@295 -- # local -ga net_devs 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@296 -- # e810=() 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@296 -- # local -ga e810 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@297 -- # x722=() 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@297 -- # local -ga x722 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@298 -- # mlx=() 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@298 -- # local -ga mlx 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:08:05.014 Found 0000:86:00.0 (0x8086 - 0x159b) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:08:05.014 Found 0000:86:00.1 (0x8086 - 0x159b) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@390 -- # [[ up == up ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:08:05.014 Found net devices under 0000:86:00.0: cvl_0_0 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@390 -- # [[ up == up ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:08:05.014 Found net devices under 0000:86:00.1: cvl_0_1 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@414 -- # is_hw=yes 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:08:05.014 17:53:57 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:08:05.014 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:08:05.014 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.201 ms 00:08:05.014 00:08:05.014 --- 10.0.0.2 ping statistics --- 00:08:05.014 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:08:05.014 rtt min/avg/max/mdev = 0.201/0.201/0.201/0.000 ms 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:08:05.014 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:08:05.014 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.112 ms 00:08:05.014 00:08:05.014 --- 10.0.0.1 ping statistics --- 00:08:05.014 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:08:05.014 rtt min/avg/max/mdev = 0.112/0.112/0.112/0.000 ms 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@422 -- # return 0 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- target/rpc.sh@24 -- # nvmfappstart -m 0xF 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@722 -- # xtrace_disable 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@481 -- # nvmfpid=460970 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@482 -- # waitforlisten 460970 00:08:05.014 17:53:58 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@829 -- # '[' -z 460970 ']' 00:08:05.015 17:53:58 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:05.015 17:53:58 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@834 -- # local max_retries=100 00:08:05.015 17:53:58 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:05.015 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:05.015 17:53:58 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@838 -- # xtrace_disable 00:08:05.015 17:53:58 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:05.015 [2024-07-15 17:53:58.288893] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:08:05.015 [2024-07-15 17:53:58.288939] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:05.015 EAL: No free 2048 kB hugepages reported on node 1 00:08:05.015 [2024-07-15 17:53:58.365429] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:08:05.015 [2024-07-15 17:53:58.445112] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:08:05.015 [2024-07-15 17:53:58.445147] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:08:05.015 [2024-07-15 17:53:58.445153] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:08:05.015 [2024-07-15 17:53:58.445159] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:08:05.015 [2024-07-15 17:53:58.445163] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:08:05.015 [2024-07-15 17:53:58.445252] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:08:05.015 [2024-07-15 17:53:58.445306] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:08:05.015 [2024-07-15 17:53:58.445388] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:08:05.015 [2024-07-15 17:53:58.445389] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:05.593 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:08:05.593 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@862 -- # return 0 00:08:05.593 17:53:59 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:08:05.593 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@728 -- # xtrace_disable 00:08:05.593 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:05.593 17:53:59 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:08:05.593 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@26 -- # rpc_cmd nvmf_get_stats 00:08:05.593 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:05.593 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:05.593 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:05.593 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@26 -- # stats='{ 00:08:05.593 "tick_rate": 2300000000, 00:08:05.593 "poll_groups": [ 00:08:05.593 { 00:08:05.593 "name": "nvmf_tgt_poll_group_000", 00:08:05.593 "admin_qpairs": 0, 00:08:05.593 "io_qpairs": 0, 00:08:05.593 "current_admin_qpairs": 0, 00:08:05.593 "current_io_qpairs": 0, 00:08:05.593 "pending_bdev_io": 0, 00:08:05.593 "completed_nvme_io": 0, 00:08:05.593 "transports": [] 00:08:05.593 }, 00:08:05.593 { 00:08:05.593 "name": "nvmf_tgt_poll_group_001", 00:08:05.593 "admin_qpairs": 0, 00:08:05.593 "io_qpairs": 0, 00:08:05.593 "current_admin_qpairs": 0, 00:08:05.593 "current_io_qpairs": 0, 00:08:05.593 "pending_bdev_io": 0, 00:08:05.593 "completed_nvme_io": 0, 00:08:05.593 "transports": [] 00:08:05.593 }, 00:08:05.593 { 00:08:05.593 "name": "nvmf_tgt_poll_group_002", 00:08:05.593 "admin_qpairs": 0, 00:08:05.593 "io_qpairs": 0, 00:08:05.593 "current_admin_qpairs": 0, 00:08:05.593 "current_io_qpairs": 0, 00:08:05.593 "pending_bdev_io": 0, 00:08:05.593 "completed_nvme_io": 0, 00:08:05.593 "transports": [] 00:08:05.593 }, 00:08:05.593 { 00:08:05.593 "name": "nvmf_tgt_poll_group_003", 00:08:05.593 "admin_qpairs": 0, 00:08:05.593 "io_qpairs": 0, 00:08:05.593 "current_admin_qpairs": 0, 00:08:05.593 "current_io_qpairs": 0, 00:08:05.593 "pending_bdev_io": 0, 00:08:05.593 "completed_nvme_io": 0, 00:08:05.593 "transports": [] 00:08:05.593 } 00:08:05.593 ] 00:08:05.594 }' 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@28 -- # jcount '.poll_groups[].name' 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@14 -- # local 'filter=.poll_groups[].name' 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@15 -- # jq '.poll_groups[].name' 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@15 -- # wc -l 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@28 -- # (( 4 == 4 )) 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@29 -- # jq '.poll_groups[0].transports[0]' 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@29 -- # [[ null == null ]] 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@31 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:05.594 [2024-07-15 17:53:59.251386] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@33 -- # rpc_cmd nvmf_get_stats 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@33 -- # stats='{ 00:08:05.594 "tick_rate": 2300000000, 00:08:05.594 "poll_groups": [ 00:08:05.594 { 00:08:05.594 "name": "nvmf_tgt_poll_group_000", 00:08:05.594 "admin_qpairs": 0, 00:08:05.594 "io_qpairs": 0, 00:08:05.594 "current_admin_qpairs": 0, 00:08:05.594 "current_io_qpairs": 0, 00:08:05.594 "pending_bdev_io": 0, 00:08:05.594 "completed_nvme_io": 0, 00:08:05.594 "transports": [ 00:08:05.594 { 00:08:05.594 "trtype": "TCP" 00:08:05.594 } 00:08:05.594 ] 00:08:05.594 }, 00:08:05.594 { 00:08:05.594 "name": "nvmf_tgt_poll_group_001", 00:08:05.594 "admin_qpairs": 0, 00:08:05.594 "io_qpairs": 0, 00:08:05.594 "current_admin_qpairs": 0, 00:08:05.594 "current_io_qpairs": 0, 00:08:05.594 "pending_bdev_io": 0, 00:08:05.594 "completed_nvme_io": 0, 00:08:05.594 "transports": [ 00:08:05.594 { 00:08:05.594 "trtype": "TCP" 00:08:05.594 } 00:08:05.594 ] 00:08:05.594 }, 00:08:05.594 { 00:08:05.594 "name": "nvmf_tgt_poll_group_002", 00:08:05.594 "admin_qpairs": 0, 00:08:05.594 "io_qpairs": 0, 00:08:05.594 "current_admin_qpairs": 0, 00:08:05.594 "current_io_qpairs": 0, 00:08:05.594 "pending_bdev_io": 0, 00:08:05.594 "completed_nvme_io": 0, 00:08:05.594 "transports": [ 00:08:05.594 { 00:08:05.594 "trtype": "TCP" 00:08:05.594 } 00:08:05.594 ] 00:08:05.594 }, 00:08:05.594 { 00:08:05.594 "name": "nvmf_tgt_poll_group_003", 00:08:05.594 "admin_qpairs": 0, 00:08:05.594 "io_qpairs": 0, 00:08:05.594 "current_admin_qpairs": 0, 00:08:05.594 "current_io_qpairs": 0, 00:08:05.594 "pending_bdev_io": 0, 00:08:05.594 "completed_nvme_io": 0, 00:08:05.594 "transports": [ 00:08:05.594 { 00:08:05.594 "trtype": "TCP" 00:08:05.594 } 00:08:05.594 ] 00:08:05.594 } 00:08:05.594 ] 00:08:05.594 }' 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@35 -- # jsum '.poll_groups[].admin_qpairs' 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].admin_qpairs' 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@20 -- # jq '.poll_groups[].admin_qpairs' 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@35 -- # (( 0 == 0 )) 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@36 -- # jsum '.poll_groups[].io_qpairs' 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].io_qpairs' 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@20 -- # jq '.poll_groups[].io_qpairs' 00:08:05.594 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@36 -- # (( 0 == 0 )) 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@38 -- # '[' rdma == tcp ']' 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@46 -- # MALLOC_BDEV_SIZE=64 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@47 -- # MALLOC_BLOCK_SIZE=512 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@49 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:05.853 Malloc1 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@52 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@53 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@54 -- # rpc_cmd nvmf_subsystem_allow_any_host -d nqn.2016-06.io.spdk:cnode1 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@55 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:05.853 [2024-07-15 17:53:59.403306] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@58 -- # NOT nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -a 10.0.0.2 -s 4420 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@648 -- # local es=0 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@650 -- # valid_exec_arg nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -a 10.0.0.2 -s 4420 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@636 -- # local arg=nvme 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@640 -- # type -t nvme 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@642 -- # type -P nvme 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@642 -- # arg=/usr/sbin/nvme 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@642 -- # [[ -x /usr/sbin/nvme ]] 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@651 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -a 10.0.0.2 -s 4420 00:08:05.853 [2024-07-15 17:53:59.431779] ctrlr.c: 822:nvmf_qpair_access_allowed: *ERROR*: Subsystem 'nqn.2016-06.io.spdk:cnode1' does not allow host 'nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562' 00:08:05.853 Failed to write to /dev/nvme-fabrics: Input/output error 00:08:05.853 could not add new controller: failed to write to nvme-fabrics device 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@651 -- # es=1 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@61 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:05.853 17:53:59 nvmf_tcp.nvmf_rpc -- target/rpc.sh@62 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:08:07.228 17:54:00 nvmf_tcp.nvmf_rpc -- target/rpc.sh@63 -- # waitforserial SPDKISFASTANDAWESOME 00:08:07.228 17:54:00 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:08:07.228 17:54:00 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:08:07.228 17:54:00 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:08:07.228 17:54:00 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- target/rpc.sh@64 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:08:09.133 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- target/rpc.sh@65 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- target/rpc.sh@68 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2016-06.io.spdk:cnode1 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- target/rpc.sh@69 -- # NOT nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@648 -- # local es=0 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@650 -- # valid_exec_arg nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@636 -- # local arg=nvme 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@640 -- # type -t nvme 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@642 -- # type -P nvme 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@642 -- # arg=/usr/sbin/nvme 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@642 -- # [[ -x /usr/sbin/nvme ]] 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@651 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:08:09.133 [2024-07-15 17:54:02.796027] ctrlr.c: 822:nvmf_qpair_access_allowed: *ERROR*: Subsystem 'nqn.2016-06.io.spdk:cnode1' does not allow host 'nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562' 00:08:09.133 Failed to write to /dev/nvme-fabrics: Input/output error 00:08:09.133 could not add new controller: failed to write to nvme-fabrics device 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@651 -- # es=1 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- target/rpc.sh@72 -- # rpc_cmd nvmf_subsystem_allow_any_host -e nqn.2016-06.io.spdk:cnode1 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:09.133 17:54:02 nvmf_tcp.nvmf_rpc -- target/rpc.sh@73 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:08:10.509 17:54:03 nvmf_tcp.nvmf_rpc -- target/rpc.sh@74 -- # waitforserial SPDKISFASTANDAWESOME 00:08:10.509 17:54:03 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:08:10.509 17:54:03 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:08:10.509 17:54:03 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:08:10.509 17:54:03 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:08:12.447 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:08:12.447 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:08:12.447 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:08:12.447 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:08:12.447 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:08:12.447 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:08:12.447 17:54:06 nvmf_tcp.nvmf_rpc -- target/rpc.sh@75 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:08:12.447 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:08:12.447 17:54:06 nvmf_tcp.nvmf_rpc -- target/rpc.sh@76 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:08:12.447 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:08:12.447 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:08:12.447 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:08:12.447 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:08:12.447 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- target/rpc.sh@78 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- target/rpc.sh@81 -- # seq 1 5 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:12.708 [2024-07-15 17:54:06.187597] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:12.708 17:54:06 nvmf_tcp.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:08:13.643 17:54:07 nvmf_tcp.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:08:13.643 17:54:07 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:08:13.644 17:54:07 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:08:13.644 17:54:07 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:08:13.644 17:54:07 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:08:16.178 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:16.178 [2024-07-15 17:54:09.432963] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:16.178 17:54:09 nvmf_tcp.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:08:17.115 17:54:10 nvmf_tcp.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:08:17.115 17:54:10 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:08:17.115 17:54:10 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:08:17.115 17:54:10 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:08:17.115 17:54:10 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:08:19.020 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:19.020 [2024-07-15 17:54:12.672136] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:19.020 17:54:12 nvmf_tcp.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:08:20.397 17:54:13 nvmf_tcp.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:08:20.397 17:54:13 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:08:20.397 17:54:13 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:08:20.397 17:54:13 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:08:20.397 17:54:13 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:08:22.303 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:22.303 17:54:15 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:22.303 17:54:16 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:22.303 17:54:16 nvmf_tcp.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:08:22.304 17:54:16 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:22.304 17:54:16 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:22.304 [2024-07-15 17:54:16.007572] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:08:22.304 17:54:16 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:22.304 17:54:16 nvmf_tcp.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:08:22.304 17:54:16 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:22.304 17:54:16 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:22.304 17:54:16 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:22.304 17:54:16 nvmf_tcp.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:08:22.304 17:54:16 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:22.304 17:54:16 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:22.304 17:54:16 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:22.304 17:54:16 nvmf_tcp.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:08:23.734 17:54:17 nvmf_tcp.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:08:23.734 17:54:17 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:08:23.734 17:54:17 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:08:23.734 17:54:17 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:08:23.734 17:54:17 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:08:25.638 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- target/rpc.sh@81 -- # for i in $(seq 1 $loops) 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- target/rpc.sh@82 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- target/rpc.sh@83 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:25.638 [2024-07-15 17:54:19.339786] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- target/rpc.sh@84 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 5 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- target/rpc.sh@85 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:25.638 17:54:19 nvmf_tcp.nvmf_rpc -- target/rpc.sh@86 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:08:27.039 17:54:20 nvmf_tcp.nvmf_rpc -- target/rpc.sh@88 -- # waitforserial SPDKISFASTANDAWESOME 00:08:27.039 17:54:20 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1198 -- # local i=0 00:08:27.039 17:54:20 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:08:27.039 17:54:20 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:08:27.039 17:54:20 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1205 -- # sleep 2 00:08:28.946 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:08:28.946 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:08:28.946 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:08:28.946 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:08:28.946 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:08:28.946 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1208 -- # return 0 00:08:28.946 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@90 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:08:28.946 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:08:28.946 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@91 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:08:28.946 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1219 -- # local i=0 00:08:28.946 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:08:28.946 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:08:28.946 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:08:28.946 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:08:28.946 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1231 -- # return 0 00:08:28.946 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@93 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:08:28.946 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:28.946 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:28.946 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:28.947 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@94 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:08:28.947 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:28.947 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:28.947 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:28.947 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@99 -- # seq 1 5 00:08:28.947 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:08:28.947 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:08:28.947 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:28.947 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:28.947 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 [2024-07-15 17:54:22.677551] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 [2024-07-15 17:54:22.725669] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 [2024-07-15 17:54:22.777810] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 [2024-07-15 17:54:22.825980] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@99 -- # for i in $(seq 1 $loops) 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@100 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDKISFASTANDAWESOME 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@101 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 [2024-07-15 17:54:22.874159] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@102 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@103 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode1 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@105 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@107 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@110 -- # rpc_cmd nvmf_get_stats 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:29.207 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@110 -- # stats='{ 00:08:29.207 "tick_rate": 2300000000, 00:08:29.208 "poll_groups": [ 00:08:29.208 { 00:08:29.208 "name": "nvmf_tgt_poll_group_000", 00:08:29.208 "admin_qpairs": 2, 00:08:29.208 "io_qpairs": 168, 00:08:29.208 "current_admin_qpairs": 0, 00:08:29.208 "current_io_qpairs": 0, 00:08:29.208 "pending_bdev_io": 0, 00:08:29.208 "completed_nvme_io": 256, 00:08:29.208 "transports": [ 00:08:29.208 { 00:08:29.208 "trtype": "TCP" 00:08:29.208 } 00:08:29.208 ] 00:08:29.208 }, 00:08:29.208 { 00:08:29.208 "name": "nvmf_tgt_poll_group_001", 00:08:29.208 "admin_qpairs": 2, 00:08:29.208 "io_qpairs": 168, 00:08:29.208 "current_admin_qpairs": 0, 00:08:29.208 "current_io_qpairs": 0, 00:08:29.208 "pending_bdev_io": 0, 00:08:29.208 "completed_nvme_io": 318, 00:08:29.208 "transports": [ 00:08:29.208 { 00:08:29.208 "trtype": "TCP" 00:08:29.208 } 00:08:29.208 ] 00:08:29.208 }, 00:08:29.208 { 00:08:29.208 "name": "nvmf_tgt_poll_group_002", 00:08:29.208 "admin_qpairs": 1, 00:08:29.208 "io_qpairs": 168, 00:08:29.208 "current_admin_qpairs": 0, 00:08:29.208 "current_io_qpairs": 0, 00:08:29.208 "pending_bdev_io": 0, 00:08:29.208 "completed_nvme_io": 217, 00:08:29.208 "transports": [ 00:08:29.208 { 00:08:29.208 "trtype": "TCP" 00:08:29.208 } 00:08:29.208 ] 00:08:29.208 }, 00:08:29.208 { 00:08:29.208 "name": "nvmf_tgt_poll_group_003", 00:08:29.208 "admin_qpairs": 2, 00:08:29.208 "io_qpairs": 168, 00:08:29.208 "current_admin_qpairs": 0, 00:08:29.208 "current_io_qpairs": 0, 00:08:29.208 "pending_bdev_io": 0, 00:08:29.208 "completed_nvme_io": 231, 00:08:29.208 "transports": [ 00:08:29.208 { 00:08:29.208 "trtype": "TCP" 00:08:29.208 } 00:08:29.208 ] 00:08:29.208 } 00:08:29.208 ] 00:08:29.208 }' 00:08:29.208 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@112 -- # jsum '.poll_groups[].admin_qpairs' 00:08:29.208 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].admin_qpairs' 00:08:29.467 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@20 -- # jq '.poll_groups[].admin_qpairs' 00:08:29.467 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:08:29.467 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@112 -- # (( 7 > 0 )) 00:08:29.467 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@113 -- # jsum '.poll_groups[].io_qpairs' 00:08:29.467 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@19 -- # local 'filter=.poll_groups[].io_qpairs' 00:08:29.467 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@20 -- # jq '.poll_groups[].io_qpairs' 00:08:29.467 17:54:22 nvmf_tcp.nvmf_rpc -- target/rpc.sh@20 -- # awk '{s+=$1}END{print s}' 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- target/rpc.sh@113 -- # (( 672 > 0 )) 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- target/rpc.sh@115 -- # '[' rdma == tcp ']' 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- target/rpc.sh@121 -- # trap - SIGINT SIGTERM EXIT 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- target/rpc.sh@123 -- # nvmftestfini 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@488 -- # nvmfcleanup 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@117 -- # sync 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@120 -- # set +e 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@121 -- # for i in {1..20} 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:08:29.467 rmmod nvme_tcp 00:08:29.467 rmmod nvme_fabrics 00:08:29.467 rmmod nvme_keyring 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@124 -- # set -e 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@125 -- # return 0 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@489 -- # '[' -n 460970 ']' 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@490 -- # killprocess 460970 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@948 -- # '[' -z 460970 ']' 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@952 -- # kill -0 460970 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@953 -- # uname 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 460970 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@966 -- # echo 'killing process with pid 460970' 00:08:29.467 killing process with pid 460970 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@967 -- # kill 460970 00:08:29.467 17:54:23 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@972 -- # wait 460970 00:08:29.726 17:54:23 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:08:29.726 17:54:23 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:08:29.726 17:54:23 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:08:29.726 17:54:23 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:08:29.726 17:54:23 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@278 -- # remove_spdk_ns 00:08:29.726 17:54:23 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:29.726 17:54:23 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:08:29.726 17:54:23 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:32.303 17:54:25 nvmf_tcp.nvmf_rpc -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:08:32.303 00:08:32.303 real 0m32.411s 00:08:32.303 user 1m40.798s 00:08:32.303 sys 0m5.575s 00:08:32.303 17:54:25 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:32.303 17:54:25 nvmf_tcp.nvmf_rpc -- common/autotest_common.sh@10 -- # set +x 00:08:32.303 ************************************ 00:08:32.303 END TEST nvmf_rpc 00:08:32.303 ************************************ 00:08:32.303 17:54:25 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:08:32.303 17:54:25 nvmf_tcp -- nvmf/nvmf.sh@30 -- # run_test nvmf_invalid /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/invalid.sh --transport=tcp 00:08:32.303 17:54:25 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:08:32.303 17:54:25 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:32.303 17:54:25 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:08:32.303 ************************************ 00:08:32.303 START TEST nvmf_invalid 00:08:32.303 ************************************ 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/invalid.sh --transport=tcp 00:08:32.303 * Looking for test storage... 00:08:32.303 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- target/invalid.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@7 -- # uname -s 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- paths/export.sh@5 -- # export PATH 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@47 -- # : 0 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@51 -- # have_pci_nics=0 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- target/invalid.sh@11 -- # multi_target_rpc=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multitarget_rpc.py 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- target/invalid.sh@12 -- # rpc=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- target/invalid.sh@13 -- # nqn=nqn.2016-06.io.spdk:cnode 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- target/invalid.sh@14 -- # target=foobar 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- target/invalid.sh@16 -- # RANDOM=0 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- target/invalid.sh@34 -- # nvmftestinit 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@448 -- # prepare_net_devs 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@410 -- # local -g is_hw=no 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@412 -- # remove_spdk_ns 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@285 -- # xtrace_disable 00:08:32.303 17:54:25 nvmf_tcp.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@291 -- # pci_devs=() 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@291 -- # local -a pci_devs 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@292 -- # pci_net_devs=() 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@293 -- # pci_drivers=() 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@293 -- # local -A pci_drivers 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@295 -- # net_devs=() 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@295 -- # local -ga net_devs 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@296 -- # e810=() 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@296 -- # local -ga e810 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@297 -- # x722=() 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@297 -- # local -ga x722 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@298 -- # mlx=() 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@298 -- # local -ga mlx 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:08:37.585 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:08:37.586 Found 0000:86:00.0 (0x8086 - 0x159b) 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:08:37.586 Found 0000:86:00.1 (0x8086 - 0x159b) 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@390 -- # [[ up == up ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:08:37.586 Found net devices under 0000:86:00.0: cvl_0_0 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@390 -- # [[ up == up ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:08:37.586 Found net devices under 0000:86:00.1: cvl_0_1 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@414 -- # is_hw=yes 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:08:37.586 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:08:37.586 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.197 ms 00:08:37.586 00:08:37.586 --- 10.0.0.2 ping statistics --- 00:08:37.586 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:08:37.586 rtt min/avg/max/mdev = 0.197/0.197/0.197/0.000 ms 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:08:37.586 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:08:37.586 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.253 ms 00:08:37.586 00:08:37.586 --- 10.0.0.1 ping statistics --- 00:08:37.586 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:08:37.586 rtt min/avg/max/mdev = 0.253/0.253/0.253/0.000 ms 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@422 -- # return 0 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- target/invalid.sh@35 -- # nvmfappstart -m 0xF 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- common/autotest_common.sh@722 -- # xtrace_disable 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@481 -- # nvmfpid=469197 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@482 -- # waitforlisten 469197 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- common/autotest_common.sh@829 -- # '[' -z 469197 ']' 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- common/autotest_common.sh@834 -- # local max_retries=100 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:37.586 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- common/autotest_common.sh@838 -- # xtrace_disable 00:08:37.586 17:54:30 nvmf_tcp.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:08:37.586 [2024-07-15 17:54:30.948960] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:08:37.586 [2024-07-15 17:54:30.949003] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:37.586 EAL: No free 2048 kB hugepages reported on node 1 00:08:37.586 [2024-07-15 17:54:31.006494] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:08:37.586 [2024-07-15 17:54:31.087495] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:08:37.586 [2024-07-15 17:54:31.087531] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:08:37.586 [2024-07-15 17:54:31.087538] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:08:37.586 [2024-07-15 17:54:31.087544] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:08:37.586 [2024-07-15 17:54:31.087549] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:08:37.586 [2024-07-15 17:54:31.087591] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:08:37.586 [2024-07-15 17:54:31.087686] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:08:37.586 [2024-07-15 17:54:31.087770] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:08:37.586 [2024-07-15 17:54:31.087771] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:08:38.156 17:54:31 nvmf_tcp.nvmf_invalid -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:08:38.156 17:54:31 nvmf_tcp.nvmf_invalid -- common/autotest_common.sh@862 -- # return 0 00:08:38.156 17:54:31 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:08:38.156 17:54:31 nvmf_tcp.nvmf_invalid -- common/autotest_common.sh@728 -- # xtrace_disable 00:08:38.156 17:54:31 nvmf_tcp.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:08:38.156 17:54:31 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:08:38.156 17:54:31 nvmf_tcp.nvmf_invalid -- target/invalid.sh@37 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini $1; exit 1' SIGINT SIGTERM EXIT 00:08:38.156 17:54:31 nvmf_tcp.nvmf_invalid -- target/invalid.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -t foobar nqn.2016-06.io.spdk:cnode16996 00:08:38.415 [2024-07-15 17:54:31.975692] nvmf_rpc.c: 396:rpc_nvmf_create_subsystem: *ERROR*: Unable to find target foobar 00:08:38.415 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@40 -- # out='request: 00:08:38.415 { 00:08:38.415 "nqn": "nqn.2016-06.io.spdk:cnode16996", 00:08:38.415 "tgt_name": "foobar", 00:08:38.415 "method": "nvmf_create_subsystem", 00:08:38.415 "req_id": 1 00:08:38.415 } 00:08:38.415 Got JSON-RPC error response 00:08:38.415 response: 00:08:38.415 { 00:08:38.415 "code": -32603, 00:08:38.415 "message": "Unable to find target foobar" 00:08:38.415 }' 00:08:38.415 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@41 -- # [[ request: 00:08:38.415 { 00:08:38.415 "nqn": "nqn.2016-06.io.spdk:cnode16996", 00:08:38.415 "tgt_name": "foobar", 00:08:38.415 "method": "nvmf_create_subsystem", 00:08:38.415 "req_id": 1 00:08:38.415 } 00:08:38.415 Got JSON-RPC error response 00:08:38.415 response: 00:08:38.415 { 00:08:38.415 "code": -32603, 00:08:38.415 "message": "Unable to find target foobar" 00:08:38.415 } == *\U\n\a\b\l\e\ \t\o\ \f\i\n\d\ \t\a\r\g\e\t* ]] 00:08:38.415 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@45 -- # echo -e '\x1f' 00:08:38.415 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -s $'SPDKISFASTANDAWESOME\037' nqn.2016-06.io.spdk:cnode29066 00:08:38.674 [2024-07-15 17:54:32.164363] nvmf_rpc.c: 413:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode29066: invalid serial number 'SPDKISFASTANDAWESOME' 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@45 -- # out='request: 00:08:38.674 { 00:08:38.674 "nqn": "nqn.2016-06.io.spdk:cnode29066", 00:08:38.674 "serial_number": "SPDKISFASTANDAWESOME\u001f", 00:08:38.674 "method": "nvmf_create_subsystem", 00:08:38.674 "req_id": 1 00:08:38.674 } 00:08:38.674 Got JSON-RPC error response 00:08:38.674 response: 00:08:38.674 { 00:08:38.674 "code": -32602, 00:08:38.674 "message": "Invalid SN SPDKISFASTANDAWESOME\u001f" 00:08:38.674 }' 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@46 -- # [[ request: 00:08:38.674 { 00:08:38.674 "nqn": "nqn.2016-06.io.spdk:cnode29066", 00:08:38.674 "serial_number": "SPDKISFASTANDAWESOME\u001f", 00:08:38.674 "method": "nvmf_create_subsystem", 00:08:38.674 "req_id": 1 00:08:38.674 } 00:08:38.674 Got JSON-RPC error response 00:08:38.674 response: 00:08:38.674 { 00:08:38.674 "code": -32602, 00:08:38.674 "message": "Invalid SN SPDKISFASTANDAWESOME\u001f" 00:08:38.674 } == *\I\n\v\a\l\i\d\ \S\N* ]] 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@50 -- # echo -e '\x1f' 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -d $'SPDK_Controller\037' nqn.2016-06.io.spdk:cnode20394 00:08:38.674 [2024-07-15 17:54:32.348957] nvmf_rpc.c: 422:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode20394: invalid model number 'SPDK_Controller' 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@50 -- # out='request: 00:08:38.674 { 00:08:38.674 "nqn": "nqn.2016-06.io.spdk:cnode20394", 00:08:38.674 "model_number": "SPDK_Controller\u001f", 00:08:38.674 "method": "nvmf_create_subsystem", 00:08:38.674 "req_id": 1 00:08:38.674 } 00:08:38.674 Got JSON-RPC error response 00:08:38.674 response: 00:08:38.674 { 00:08:38.674 "code": -32602, 00:08:38.674 "message": "Invalid MN SPDK_Controller\u001f" 00:08:38.674 }' 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@51 -- # [[ request: 00:08:38.674 { 00:08:38.674 "nqn": "nqn.2016-06.io.spdk:cnode20394", 00:08:38.674 "model_number": "SPDK_Controller\u001f", 00:08:38.674 "method": "nvmf_create_subsystem", 00:08:38.674 "req_id": 1 00:08:38.674 } 00:08:38.674 Got JSON-RPC error response 00:08:38.674 response: 00:08:38.674 { 00:08:38.674 "code": -32602, 00:08:38.674 "message": "Invalid MN SPDK_Controller\u001f" 00:08:38.674 } == *\I\n\v\a\l\i\d\ \M\N* ]] 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@54 -- # gen_random_s 21 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@19 -- # local length=21 ll 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@21 -- # chars=('32' '33' '34' '35' '36' '37' '38' '39' '40' '41' '42' '43' '44' '45' '46' '47' '48' '49' '50' '51' '52' '53' '54' '55' '56' '57' '58' '59' '60' '61' '62' '63' '64' '65' '66' '67' '68' '69' '70' '71' '72' '73' '74' '75' '76' '77' '78' '79' '80' '81' '82' '83' '84' '85' '86' '87' '88' '89' '90' '91' '92' '93' '94' '95' '96' '97' '98' '99' '100' '101' '102' '103' '104' '105' '106' '107' '108' '109' '110' '111' '112' '113' '114' '115' '116' '117' '118' '119' '120' '121' '122' '123' '124' '125' '126' '127') 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@21 -- # local chars 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@22 -- # local string 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll = 0 )) 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 47 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2f' 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=/ 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 102 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x66' 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=f 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 37 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x25' 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=% 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.674 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 102 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x66' 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=f 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 78 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x4e' 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=N 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 73 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x49' 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=I 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 92 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x5c' 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+='\' 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 89 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x59' 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=Y 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 115 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x73' 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=s 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 41 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x29' 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=')' 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 77 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x4d' 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=M 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 119 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x77' 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=w 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 101 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x65' 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=e 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 95 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x5f' 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=_ 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 109 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x6d' 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=m 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 45 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2d' 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=- 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 97 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x61' 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=a 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 34 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x22' 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+='"' 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 34 00:08:38.934 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x22' 00:08:38.935 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+='"' 00:08:38.935 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.935 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.935 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 91 00:08:38.935 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x5b' 00:08:38.935 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+='[' 00:08:38.935 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.935 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.935 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 105 00:08:38.935 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x69' 00:08:38.935 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=i 00:08:38.935 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:38.935 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:38.935 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@28 -- # [[ / == \- ]] 00:08:38.935 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@31 -- # echo '/f%fNI\Ys)Mwe_m-a""[i' 00:08:38.935 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem -s '/f%fNI\Ys)Mwe_m-a""[i' nqn.2016-06.io.spdk:cnode15725 00:08:39.195 [2024-07-15 17:54:32.666053] nvmf_rpc.c: 413:rpc_nvmf_create_subsystem: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode15725: invalid serial number '/f%fNI\Ys)Mwe_m-a""[i' 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@54 -- # out='request: 00:08:39.195 { 00:08:39.195 "nqn": "nqn.2016-06.io.spdk:cnode15725", 00:08:39.195 "serial_number": "/f%fNI\\Ys)Mwe_m-a\"\"[i", 00:08:39.195 "method": "nvmf_create_subsystem", 00:08:39.195 "req_id": 1 00:08:39.195 } 00:08:39.195 Got JSON-RPC error response 00:08:39.195 response: 00:08:39.195 { 00:08:39.195 "code": -32602, 00:08:39.195 "message": "Invalid SN /f%fNI\\Ys)Mwe_m-a\"\"[i" 00:08:39.195 }' 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@55 -- # [[ request: 00:08:39.195 { 00:08:39.195 "nqn": "nqn.2016-06.io.spdk:cnode15725", 00:08:39.195 "serial_number": "/f%fNI\\Ys)Mwe_m-a\"\"[i", 00:08:39.195 "method": "nvmf_create_subsystem", 00:08:39.195 "req_id": 1 00:08:39.195 } 00:08:39.195 Got JSON-RPC error response 00:08:39.195 response: 00:08:39.195 { 00:08:39.195 "code": -32602, 00:08:39.195 "message": "Invalid SN /f%fNI\\Ys)Mwe_m-a\"\"[i" 00:08:39.195 } == *\I\n\v\a\l\i\d\ \S\N* ]] 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@58 -- # gen_random_s 41 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@19 -- # local length=41 ll 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@21 -- # chars=('32' '33' '34' '35' '36' '37' '38' '39' '40' '41' '42' '43' '44' '45' '46' '47' '48' '49' '50' '51' '52' '53' '54' '55' '56' '57' '58' '59' '60' '61' '62' '63' '64' '65' '66' '67' '68' '69' '70' '71' '72' '73' '74' '75' '76' '77' '78' '79' '80' '81' '82' '83' '84' '85' '86' '87' '88' '89' '90' '91' '92' '93' '94' '95' '96' '97' '98' '99' '100' '101' '102' '103' '104' '105' '106' '107' '108' '109' '110' '111' '112' '113' '114' '115' '116' '117' '118' '119' '120' '121' '122' '123' '124' '125' '126' '127') 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@21 -- # local chars 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@22 -- # local string 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll = 0 )) 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 40 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x28' 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+='(' 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 104 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x68' 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=h 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 42 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2a' 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+='*' 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 57 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x39' 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=9 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 86 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x56' 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=V 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 117 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x75' 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=u 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 120 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x78' 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=x 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 127 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x7f' 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=$'\177' 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 62 00:08:39.195 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x3e' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+='>' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 47 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2f' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=/ 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 47 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2f' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=/ 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 103 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x67' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=g 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 83 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x53' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=S 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 60 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x3c' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+='<' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 94 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x5e' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+='^' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 60 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x3c' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+='<' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 65 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x41' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=A 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 44 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x2c' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=, 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 93 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x5d' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=']' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 60 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x3c' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+='<' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 52 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x34' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=4 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 109 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x6d' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=m 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 121 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x79' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=y 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 86 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x56' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=V 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 112 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x70' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=p 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 94 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x5e' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+='^' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 40 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x28' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+='(' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 99 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x63' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=c 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 100 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x64' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=d 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 66 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x42' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=B 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 91 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x5b' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+='[' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 81 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x51' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=Q 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 68 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x44' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=D 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 54 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x36' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=6 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 37 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x25' 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=% 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.196 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 90 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x5a' 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=Z 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 99 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x63' 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=c 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 121 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x79' 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=y 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 54 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x36' 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=6 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 93 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x5d' 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=']' 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # printf %x 69 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # echo -e '\x45' 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@25 -- # string+=E 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll++ )) 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@24 -- # (( ll < length )) 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@28 -- # [[ ( == \- ]] 00:08:39.457 17:54:32 nvmf_tcp.nvmf_invalid -- target/invalid.sh@31 -- # echo '(h*9Vux>//gS<^//gS<^//gS<^//gS<^//gS<^ /dev/null' 00:08:41.537 17:54:35 nvmf_tcp.nvmf_invalid -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:43.445 17:54:37 nvmf_tcp.nvmf_invalid -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:08:43.445 00:08:43.445 real 0m11.649s 00:08:43.445 user 0m19.582s 00:08:43.445 sys 0m4.968s 00:08:43.445 17:54:37 nvmf_tcp.nvmf_invalid -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:43.445 17:54:37 nvmf_tcp.nvmf_invalid -- common/autotest_common.sh@10 -- # set +x 00:08:43.445 ************************************ 00:08:43.445 END TEST nvmf_invalid 00:08:43.445 ************************************ 00:08:43.445 17:54:37 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:08:43.445 17:54:37 nvmf_tcp -- nvmf/nvmf.sh@31 -- # run_test nvmf_abort /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort.sh --transport=tcp 00:08:43.445 17:54:37 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:08:43.445 17:54:37 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:43.445 17:54:37 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:08:43.705 ************************************ 00:08:43.705 START TEST nvmf_abort 00:08:43.705 ************************************ 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort.sh --transport=tcp 00:08:43.705 * Looking for test storage... 00:08:43.705 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- target/abort.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@7 -- # uname -s 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- paths/export.sh@5 -- # export PATH 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@47 -- # : 0 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@51 -- # have_pci_nics=0 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- target/abort.sh@11 -- # MALLOC_BDEV_SIZE=64 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- target/abort.sh@12 -- # MALLOC_BLOCK_SIZE=4096 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- target/abort.sh@14 -- # nvmftestinit 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@448 -- # prepare_net_devs 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@410 -- # local -g is_hw=no 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@412 -- # remove_spdk_ns 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- nvmf/common.sh@285 -- # xtrace_disable 00:08:43.705 17:54:37 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@291 -- # pci_devs=() 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@291 -- # local -a pci_devs 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@292 -- # pci_net_devs=() 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@293 -- # pci_drivers=() 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@293 -- # local -A pci_drivers 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@295 -- # net_devs=() 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@295 -- # local -ga net_devs 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@296 -- # e810=() 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@296 -- # local -ga e810 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@297 -- # x722=() 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@297 -- # local -ga x722 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@298 -- # mlx=() 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@298 -- # local -ga mlx 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:08:48.975 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:08:48.976 Found 0000:86:00.0 (0x8086 - 0x159b) 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:08:48.976 Found 0000:86:00.1 (0x8086 - 0x159b) 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@390 -- # [[ up == up ]] 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:08:48.976 Found net devices under 0000:86:00.0: cvl_0_0 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@390 -- # [[ up == up ]] 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:08:48.976 Found net devices under 0000:86:00.1: cvl_0_1 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@414 -- # is_hw=yes 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:08:48.976 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:08:49.235 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:08:49.235 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.172 ms 00:08:49.235 00:08:49.235 --- 10.0.0.2 ping statistics --- 00:08:49.235 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:08:49.235 rtt min/avg/max/mdev = 0.172/0.172/0.172/0.000 ms 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:08:49.235 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:08:49.235 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.248 ms 00:08:49.235 00:08:49.235 --- 10.0.0.1 ping statistics --- 00:08:49.235 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:08:49.235 rtt min/avg/max/mdev = 0.248/0.248/0.248/0.000 ms 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@422 -- # return 0 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- target/abort.sh@15 -- # nvmfappstart -m 0xE 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@722 -- # xtrace_disable 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@481 -- # nvmfpid=473568 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@482 -- # waitforlisten 473568 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@829 -- # '[' -z 473568 ']' 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@834 -- # local max_retries=100 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:08:49.235 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@838 -- # xtrace_disable 00:08:49.235 17:54:42 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:08:49.235 [2024-07-15 17:54:42.806015] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:08:49.235 [2024-07-15 17:54:42.806056] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:08:49.235 EAL: No free 2048 kB hugepages reported on node 1 00:08:49.235 [2024-07-15 17:54:42.864615] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:08:49.235 [2024-07-15 17:54:42.942369] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:08:49.235 [2024-07-15 17:54:42.942407] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:08:49.235 [2024-07-15 17:54:42.942414] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:08:49.235 [2024-07-15 17:54:42.942421] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:08:49.235 [2024-07-15 17:54:42.942426] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:08:49.235 [2024-07-15 17:54:42.942549] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:08:49.235 [2024-07-15 17:54:42.942634] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:08:49.235 [2024-07-15 17:54:42.942635] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@862 -- # return 0 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@728 -- # xtrace_disable 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- target/abort.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 -a 256 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:08:50.170 [2024-07-15 17:54:43.646879] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- target/abort.sh@20 -- # rpc_cmd bdev_malloc_create 64 4096 -b Malloc0 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:08:50.170 Malloc0 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- target/abort.sh@21 -- # rpc_cmd bdev_delay_create -b Malloc0 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:08:50.170 Delay0 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- target/abort.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- target/abort.sh@25 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 Delay0 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- target/abort.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:08:50.170 [2024-07-15 17:54:43.719591] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- target/abort.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:50.170 17:54:43 nvmf_tcp.nvmf_abort -- target/abort.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -c 0x1 -t 1 -l warning -q 128 00:08:50.170 EAL: No free 2048 kB hugepages reported on node 1 00:08:50.170 [2024-07-15 17:54:43.824548] nvme_fabric.c: 295:nvme_fabric_discover_probe: *WARNING*: Skipping unsupported current discovery service or discovery service referral 00:08:52.704 Initializing NVMe Controllers 00:08:52.704 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode0 00:08:52.704 controller IO queue size 128 less than required 00:08:52.704 Consider using lower queue depth or small IO size because IO requests may be queued at the NVMe driver. 00:08:52.704 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 0 00:08:52.704 Initialization complete. Launching workers. 00:08:52.704 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 I/O completed: 123, failed: 43359 00:08:52.704 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) abort submitted 43420, failed to submit 62 00:08:52.704 success 43363, unsuccess 57, failed 0 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- target/abort.sh@34 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@559 -- # xtrace_disable 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- target/abort.sh@36 -- # trap - SIGINT SIGTERM EXIT 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- target/abort.sh@38 -- # nvmftestfini 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- nvmf/common.sh@488 -- # nvmfcleanup 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- nvmf/common.sh@117 -- # sync 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- nvmf/common.sh@120 -- # set +e 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- nvmf/common.sh@121 -- # for i in {1..20} 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:08:52.704 rmmod nvme_tcp 00:08:52.704 rmmod nvme_fabrics 00:08:52.704 rmmod nvme_keyring 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- nvmf/common.sh@124 -- # set -e 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- nvmf/common.sh@125 -- # return 0 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- nvmf/common.sh@489 -- # '[' -n 473568 ']' 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- nvmf/common.sh@490 -- # killprocess 473568 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@948 -- # '[' -z 473568 ']' 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@952 -- # kill -0 473568 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@953 -- # uname 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 473568 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@966 -- # echo 'killing process with pid 473568' 00:08:52.704 killing process with pid 473568 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@967 -- # kill 473568 00:08:52.704 17:54:45 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@972 -- # wait 473568 00:08:52.704 17:54:46 nvmf_tcp.nvmf_abort -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:08:52.704 17:54:46 nvmf_tcp.nvmf_abort -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:08:52.704 17:54:46 nvmf_tcp.nvmf_abort -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:08:52.704 17:54:46 nvmf_tcp.nvmf_abort -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:08:52.704 17:54:46 nvmf_tcp.nvmf_abort -- nvmf/common.sh@278 -- # remove_spdk_ns 00:08:52.704 17:54:46 nvmf_tcp.nvmf_abort -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:52.704 17:54:46 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:08:52.704 17:54:46 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:54.644 17:54:48 nvmf_tcp.nvmf_abort -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:08:54.644 00:08:54.644 real 0m11.066s 00:08:54.644 user 0m12.841s 00:08:54.644 sys 0m5.086s 00:08:54.644 17:54:48 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@1124 -- # xtrace_disable 00:08:54.644 17:54:48 nvmf_tcp.nvmf_abort -- common/autotest_common.sh@10 -- # set +x 00:08:54.644 ************************************ 00:08:54.644 END TEST nvmf_abort 00:08:54.644 ************************************ 00:08:54.644 17:54:48 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:08:54.644 17:54:48 nvmf_tcp -- nvmf/nvmf.sh@32 -- # run_test nvmf_ns_hotplug_stress /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh --transport=tcp 00:08:54.644 17:54:48 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:08:54.644 17:54:48 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:08:54.644 17:54:48 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:08:54.644 ************************************ 00:08:54.644 START TEST nvmf_ns_hotplug_stress 00:08:54.644 ************************************ 00:08:54.644 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh --transport=tcp 00:08:54.904 * Looking for test storage... 00:08:54.904 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@7 -- # uname -s 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- paths/export.sh@5 -- # export PATH 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@47 -- # : 0 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@51 -- # have_pci_nics=0 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@11 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@22 -- # nvmftestinit 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@448 -- # prepare_net_devs 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@410 -- # local -g is_hw=no 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@412 -- # remove_spdk_ns 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@285 -- # xtrace_disable 00:08:54.904 17:54:48 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@291 -- # pci_devs=() 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@291 -- # local -a pci_devs 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@292 -- # pci_net_devs=() 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@293 -- # pci_drivers=() 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@293 -- # local -A pci_drivers 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@295 -- # net_devs=() 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@295 -- # local -ga net_devs 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@296 -- # e810=() 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@296 -- # local -ga e810 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@297 -- # x722=() 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@297 -- # local -ga x722 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@298 -- # mlx=() 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@298 -- # local -ga mlx 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:09:00.182 Found 0000:86:00.0 (0x8086 - 0x159b) 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:09:00.182 Found 0000:86:00.1 (0x8086 - 0x159b) 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:09:00.182 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@390 -- # [[ up == up ]] 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:09:00.183 Found net devices under 0000:86:00.0: cvl_0_0 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@390 -- # [[ up == up ]] 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:09:00.183 Found net devices under 0000:86:00.1: cvl_0_1 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@414 -- # is_hw=yes 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:09:00.183 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:09:00.183 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.183 ms 00:09:00.183 00:09:00.183 --- 10.0.0.2 ping statistics --- 00:09:00.183 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:00.183 rtt min/avg/max/mdev = 0.183/0.183/0.183/0.000 ms 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:09:00.183 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:09:00.183 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.196 ms 00:09:00.183 00:09:00.183 --- 10.0.0.1 ping statistics --- 00:09:00.183 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:00.183 rtt min/avg/max/mdev = 0.196/0.196/0.196/0.000 ms 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@422 -- # return 0 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@23 -- # nvmfappstart -m 0xE 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@722 -- # xtrace_disable 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@481 -- # nvmfpid=477568 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@482 -- # waitforlisten 477568 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@829 -- # '[' -z 477568 ']' 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@834 -- # local max_retries=100 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:00.183 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@838 -- # xtrace_disable 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:09:00.183 17:54:53 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:09:00.183 [2024-07-15 17:54:53.856943] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:09:00.183 [2024-07-15 17:54:53.856986] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:00.183 EAL: No free 2048 kB hugepages reported on node 1 00:09:00.442 [2024-07-15 17:54:53.912869] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:09:00.442 [2024-07-15 17:54:53.992528] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:09:00.442 [2024-07-15 17:54:53.992564] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:09:00.442 [2024-07-15 17:54:53.992571] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:09:00.442 [2024-07-15 17:54:53.992577] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:09:00.442 [2024-07-15 17:54:53.992583] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:09:00.442 [2024-07-15 17:54:53.992681] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:09:00.442 [2024-07-15 17:54:53.992703] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:09:00.442 [2024-07-15 17:54:53.992704] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:09:01.010 17:54:54 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:09:01.010 17:54:54 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@862 -- # return 0 00:09:01.010 17:54:54 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:09:01.010 17:54:54 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@728 -- # xtrace_disable 00:09:01.010 17:54:54 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:09:01.010 17:54:54 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:09:01.010 17:54:54 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@25 -- # null_size=1000 00:09:01.010 17:54:54 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:09:01.269 [2024-07-15 17:54:54.849685] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:09:01.269 17:54:54 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:09:01.528 17:54:55 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:09:01.528 [2024-07-15 17:54:55.246711] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:09:01.787 17:54:55 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:09:01.787 17:54:55 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 512 -b Malloc0 00:09:02.046 Malloc0 00:09:02.046 17:54:55 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_delay_create -b Malloc0 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:09:02.305 Delay0 00:09:02.305 17:54:55 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:02.305 17:54:55 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create NULL1 1000 512 00:09:02.564 NULL1 00:09:02.564 17:54:56 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 NULL1 00:09:02.823 17:54:56 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@42 -- # PERF_PID=477915 00:09:02.823 17:54:56 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0x1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 30 -q 128 -w randread -o 512 -Q 1000 00:09:02.823 17:54:56 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:02.823 17:54:56 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:02.823 EAL: No free 2048 kB hugepages reported on node 1 00:09:02.823 17:54:56 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:03.083 17:54:56 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1001 00:09:03.083 17:54:56 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1001 00:09:03.342 true 00:09:03.342 17:54:56 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:03.342 17:54:56 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:03.600 17:54:57 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:03.600 17:54:57 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1002 00:09:03.600 17:54:57 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1002 00:09:03.858 true 00:09:03.858 17:54:57 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:03.858 17:54:57 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:04.116 17:54:57 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:04.374 17:54:57 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1003 00:09:04.374 17:54:57 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1003 00:09:04.374 true 00:09:04.374 17:54:58 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:04.374 17:54:58 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:04.632 17:54:58 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:04.889 17:54:58 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1004 00:09:04.889 17:54:58 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1004 00:09:05.147 true 00:09:05.147 17:54:58 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:05.147 17:54:58 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:05.147 17:54:58 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:05.405 17:54:59 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1005 00:09:05.405 17:54:59 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1005 00:09:05.664 true 00:09:05.664 17:54:59 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:05.664 17:54:59 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:05.923 17:54:59 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:05.923 17:54:59 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1006 00:09:05.923 17:54:59 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1006 00:09:06.181 true 00:09:06.181 17:54:59 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:06.181 17:54:59 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:06.439 17:54:59 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:06.698 17:55:00 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1007 00:09:06.698 17:55:00 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1007 00:09:06.698 true 00:09:06.698 17:55:00 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:06.698 17:55:00 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:06.956 17:55:00 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:07.215 17:55:00 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1008 00:09:07.215 17:55:00 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1008 00:09:07.473 true 00:09:07.473 17:55:00 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:07.473 17:55:00 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:07.473 17:55:01 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:07.732 17:55:01 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1009 00:09:07.732 17:55:01 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1009 00:09:07.991 true 00:09:07.991 17:55:01 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:07.991 17:55:01 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:08.250 17:55:01 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:08.250 17:55:01 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1010 00:09:08.250 17:55:01 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1010 00:09:08.510 true 00:09:08.510 17:55:02 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:08.510 17:55:02 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:08.770 17:55:02 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:09.030 17:55:02 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1011 00:09:09.030 17:55:02 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1011 00:09:09.030 true 00:09:09.030 17:55:02 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:09.030 17:55:02 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:09.289 17:55:02 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:09.548 17:55:03 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1012 00:09:09.549 17:55:03 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1012 00:09:09.549 true 00:09:09.807 17:55:03 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:09.807 17:55:03 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:09.807 17:55:03 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:10.066 17:55:03 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1013 00:09:10.066 17:55:03 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1013 00:09:10.325 true 00:09:10.325 17:55:03 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:10.325 17:55:03 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:10.584 17:55:04 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:10.584 17:55:04 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1014 00:09:10.584 17:55:04 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1014 00:09:10.844 true 00:09:10.844 17:55:04 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:10.844 17:55:04 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:11.113 17:55:04 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:11.113 17:55:04 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1015 00:09:11.113 17:55:04 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1015 00:09:11.398 true 00:09:11.398 17:55:05 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:11.398 17:55:05 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:11.657 17:55:05 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:11.916 17:55:05 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1016 00:09:11.916 17:55:05 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1016 00:09:11.916 true 00:09:11.916 17:55:05 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:11.916 17:55:05 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:12.175 17:55:05 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:12.434 17:55:05 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1017 00:09:12.434 17:55:05 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1017 00:09:12.434 true 00:09:12.434 17:55:06 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:12.434 17:55:06 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:12.693 17:55:06 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:12.952 17:55:06 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1018 00:09:12.952 17:55:06 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1018 00:09:13.211 true 00:09:13.211 17:55:06 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:13.211 17:55:06 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:13.211 17:55:06 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:13.470 17:55:07 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1019 00:09:13.470 17:55:07 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1019 00:09:13.729 true 00:09:13.729 17:55:07 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:13.729 17:55:07 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:13.989 17:55:07 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:13.989 17:55:07 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1020 00:09:13.989 17:55:07 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1020 00:09:14.247 true 00:09:14.247 17:55:07 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:14.247 17:55:07 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:14.507 17:55:08 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:14.766 17:55:08 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1021 00:09:14.766 17:55:08 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1021 00:09:14.766 true 00:09:14.766 17:55:08 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:14.766 17:55:08 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:15.025 17:55:08 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:15.284 17:55:08 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1022 00:09:15.284 17:55:08 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1022 00:09:15.543 true 00:09:15.543 17:55:09 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:15.543 17:55:09 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:15.543 17:55:09 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:15.802 17:55:09 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1023 00:09:15.802 17:55:09 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1023 00:09:16.062 true 00:09:16.062 17:55:09 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:16.062 17:55:09 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:16.322 17:55:09 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:16.322 17:55:10 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1024 00:09:16.322 17:55:10 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1024 00:09:16.580 true 00:09:16.580 17:55:10 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:16.580 17:55:10 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:16.839 17:55:10 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:17.097 17:55:10 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1025 00:09:17.097 17:55:10 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1025 00:09:17.097 true 00:09:17.097 17:55:10 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:17.097 17:55:10 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:17.356 17:55:10 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:17.615 17:55:11 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1026 00:09:17.615 17:55:11 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1026 00:09:17.874 true 00:09:17.874 17:55:11 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:17.874 17:55:11 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:17.874 17:55:11 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:18.134 17:55:11 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1027 00:09:18.134 17:55:11 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1027 00:09:18.393 true 00:09:18.393 17:55:11 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:18.393 17:55:11 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:18.668 17:55:12 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:18.668 17:55:12 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1028 00:09:18.668 17:55:12 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1028 00:09:18.927 true 00:09:18.927 17:55:12 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:18.927 17:55:12 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:19.186 17:55:12 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:19.447 17:55:12 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1029 00:09:19.447 17:55:12 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1029 00:09:19.447 true 00:09:19.447 17:55:13 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:19.447 17:55:13 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:19.706 17:55:13 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:19.965 17:55:13 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1030 00:09:19.965 17:55:13 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1030 00:09:19.965 true 00:09:20.224 17:55:13 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:20.224 17:55:13 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:20.224 17:55:13 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:20.483 17:55:14 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1031 00:09:20.483 17:55:14 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1031 00:09:20.741 true 00:09:20.741 17:55:14 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:20.741 17:55:14 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:21.000 17:55:14 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:21.000 17:55:14 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1032 00:09:21.000 17:55:14 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1032 00:09:21.260 true 00:09:21.260 17:55:14 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:21.260 17:55:14 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:21.520 17:55:15 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:21.779 17:55:15 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1033 00:09:21.779 17:55:15 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1033 00:09:21.779 true 00:09:21.779 17:55:15 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:21.779 17:55:15 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:22.038 17:55:15 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:22.297 17:55:15 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1034 00:09:22.297 17:55:15 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1034 00:09:22.297 true 00:09:22.557 17:55:16 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:22.557 17:55:16 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:22.557 17:55:16 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:22.815 17:55:16 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1035 00:09:22.815 17:55:16 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1035 00:09:23.075 true 00:09:23.075 17:55:16 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:23.075 17:55:16 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:23.335 17:55:16 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:23.335 17:55:17 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1036 00:09:23.335 17:55:17 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1036 00:09:23.595 true 00:09:23.595 17:55:17 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:23.595 17:55:17 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:23.854 17:55:17 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:24.111 17:55:17 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1037 00:09:24.112 17:55:17 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1037 00:09:24.112 true 00:09:24.112 17:55:17 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:24.112 17:55:17 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:24.370 17:55:17 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:24.628 17:55:18 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1038 00:09:24.628 17:55:18 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1038 00:09:24.628 true 00:09:24.886 17:55:18 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:24.886 17:55:18 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:24.886 17:55:18 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:25.169 17:55:18 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1039 00:09:25.169 17:55:18 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1039 00:09:25.428 true 00:09:25.428 17:55:18 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:25.428 17:55:18 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:25.686 17:55:19 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:25.686 17:55:19 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1040 00:09:25.686 17:55:19 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1040 00:09:25.943 true 00:09:25.943 17:55:19 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:25.943 17:55:19 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:26.201 17:55:19 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:26.201 17:55:19 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1041 00:09:26.201 17:55:19 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1041 00:09:26.459 true 00:09:26.459 17:55:20 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:26.459 17:55:20 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:26.718 17:55:20 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:26.976 17:55:20 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1042 00:09:26.976 17:55:20 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1042 00:09:26.976 true 00:09:26.976 17:55:20 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:26.976 17:55:20 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:27.235 17:55:20 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:27.493 17:55:21 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1043 00:09:27.493 17:55:21 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1043 00:09:27.752 true 00:09:27.752 17:55:21 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:27.752 17:55:21 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:27.752 17:55:21 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:28.011 17:55:21 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1044 00:09:28.011 17:55:21 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1044 00:09:28.294 true 00:09:28.294 17:55:21 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:28.294 17:55:21 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:28.553 17:55:22 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:28.553 17:55:22 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1045 00:09:28.553 17:55:22 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1045 00:09:28.811 true 00:09:28.811 17:55:22 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:28.811 17:55:22 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:29.070 17:55:22 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:29.329 17:55:22 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1046 00:09:29.329 17:55:22 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1046 00:09:29.329 true 00:09:29.329 17:55:23 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:29.329 17:55:23 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:29.588 17:55:23 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:29.847 17:55:23 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1047 00:09:29.847 17:55:23 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1047 00:09:29.847 true 00:09:30.105 17:55:23 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:30.105 17:55:23 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:30.105 17:55:23 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:30.365 17:55:23 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1048 00:09:30.365 17:55:23 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1048 00:09:30.635 true 00:09:30.635 17:55:24 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:30.635 17:55:24 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:30.902 17:55:24 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:30.902 17:55:24 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1049 00:09:30.902 17:55:24 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1049 00:09:31.162 true 00:09:31.162 17:55:24 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:31.162 17:55:24 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:31.422 17:55:24 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:31.422 17:55:25 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1050 00:09:31.422 17:55:25 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1050 00:09:31.682 true 00:09:31.682 17:55:25 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:31.682 17:55:25 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:31.941 17:55:25 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:32.200 17:55:25 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1051 00:09:32.200 17:55:25 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1051 00:09:32.200 true 00:09:32.200 17:55:25 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:32.200 17:55:25 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:32.460 17:55:26 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:32.719 17:55:26 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1052 00:09:32.719 17:55:26 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1052 00:09:32.978 true 00:09:32.978 17:55:26 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:32.978 17:55:26 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:32.978 Initializing NVMe Controllers 00:09:32.978 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:09:32.978 Controller SPDK bdev Controller (SPDK00000000000001 ): Skipping inactive NS 1 00:09:32.978 Controller IO queue size 128, less than required. 00:09:32.978 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:09:32.978 WARNING: Some requested NVMe devices were skipped 00:09:32.978 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:09:32.978 Initialization complete. Launching workers. 00:09:32.978 ======================================================== 00:09:32.978 Latency(us) 00:09:32.978 Device Information : IOPS MiB/s Average min max 00:09:32.979 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 27228.37 13.30 4700.90 2246.66 8448.20 00:09:32.979 ======================================================== 00:09:32.979 Total : 27228.37 13.30 4700.90 2246.66 8448.20 00:09:32.979 00:09:32.979 17:55:26 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:09:33.237 17:55:26 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@49 -- # null_size=1053 00:09:33.237 17:55:26 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_resize NULL1 1053 00:09:33.497 true 00:09:33.497 17:55:27 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@44 -- # kill -0 477915 00:09:33.497 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/ns_hotplug_stress.sh: line 44: kill: (477915) - No such process 00:09:33.497 17:55:27 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@53 -- # wait 477915 00:09:33.497 17:55:27 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:33.756 17:55:27 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:33.756 17:55:27 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@58 -- # nthreads=8 00:09:33.756 17:55:27 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@58 -- # pids=() 00:09:33.756 17:55:27 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i = 0 )) 00:09:33.756 17:55:27 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:09:33.756 17:55:27 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null0 100 4096 00:09:34.015 null0 00:09:34.015 17:55:27 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:09:34.015 17:55:27 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:09:34.015 17:55:27 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null1 100 4096 00:09:34.273 null1 00:09:34.274 17:55:27 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:09:34.274 17:55:27 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:09:34.274 17:55:27 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null2 100 4096 00:09:34.274 null2 00:09:34.274 17:55:27 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:09:34.274 17:55:27 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:09:34.274 17:55:27 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null3 100 4096 00:09:34.532 null3 00:09:34.532 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:09:34.532 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:09:34.532 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null4 100 4096 00:09:34.792 null4 00:09:34.792 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:09:34.792 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:09:34.792 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null5 100 4096 00:09:34.792 null5 00:09:34.792 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:09:34.792 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:09:35.052 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null6 100 4096 00:09:35.052 null6 00:09:35.052 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:09:35.052 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:09:35.052 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_null_create null7 100 4096 00:09:35.312 null7 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( ++i )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@59 -- # (( i < nthreads )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i = 0 )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 1 null0 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=1 bdev=null0 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 2 null1 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=2 bdev=null1 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 3 null2 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=3 bdev=null2 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 4 null3 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=4 bdev=null3 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 5 null4 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=5 bdev=null4 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 6 null5 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=6 bdev=null5 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 7 null6 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=7 bdev=null6 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@64 -- # pids+=($!) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( ++i )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@62 -- # (( i < nthreads )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@63 -- # add_remove 8 null7 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@66 -- # wait 483501 483502 483504 483506 483508 483510 483512 483514 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@14 -- # local nsid=8 bdev=null7 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i = 0 )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:35.312 17:55:28 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:35.572 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:09:35.831 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:35.831 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:09:35.831 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:35.831 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:09:35.831 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:09:35.831 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:09:35.831 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:09:35.831 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:09:36.090 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.091 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:09:36.350 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:09:36.350 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:36.350 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:09:36.350 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:09:36.350 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:09:36.350 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:09:36.350 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:09:36.350 17:55:29 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.350 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:09:36.609 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:36.609 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:36.609 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:09:36.609 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:09:36.609 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:09:36.609 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:09:36.609 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:09:36.609 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:36.869 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.128 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:09:37.387 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:09:37.387 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:09:37.387 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:37.387 17:55:30 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:09:37.387 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:09:37.387 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:09:37.387 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:09:37.387 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:37.646 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.646 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.646 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:09:37.646 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.646 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.646 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:09:37.646 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.646 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.646 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.646 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:09:37.646 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.646 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:09:37.646 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.646 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.646 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:09:37.646 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.646 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.646 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:09:37.646 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.646 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.647 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:09:37.647 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.647 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.647 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:09:37.647 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:09:37.647 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:37.906 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:09:38.166 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:09:38.166 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:38.166 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:09:38.166 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:09:38.166 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:09:38.166 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:09:38.167 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:09:38.167 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:38.425 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:38.425 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:38.425 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:09:38.425 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:38.425 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:38.425 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:09:38.425 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:38.425 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:38.425 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:09:38.425 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:38.425 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:38.425 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:09:38.425 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:38.425 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:38.425 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:09:38.425 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:38.425 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:38.426 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:09:38.426 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:38.426 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:38.426 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:09:38.426 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:38.426 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:38.426 17:55:31 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:09:38.426 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:09:38.426 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:09:38.426 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:09:38.426 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:38.426 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:09:38.426 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:38.426 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:09:38.426 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 3 nqn.2016-06.io.spdk:cnode1 null2 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 5 nqn.2016-06.io.spdk:cnode1 null4 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 4 nqn.2016-06.io.spdk:cnode1 null3 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 7 nqn.2016-06.io.spdk:cnode1 null6 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 6 nqn.2016-06.io.spdk:cnode1 null5 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 2 nqn.2016-06.io.spdk:cnode1 null1 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 8 nqn.2016-06.io.spdk:cnode1 null7 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:38.710 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@17 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns -n 1 nqn.2016-06.io.spdk:cnode1 null0 00:09:38.978 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 3 00:09:38.978 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 8 00:09:38.978 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 5 00:09:38.978 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 4 00:09:38.978 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:09:38.978 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 7 00:09:38.978 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 6 00:09:38.978 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:09:38.978 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:38.978 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:38.978 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:38.978 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:38.978 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:38.978 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:38.978 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:38.978 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( ++i )) 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@16 -- # (( i < 10 )) 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@68 -- # trap - SIGINT SIGTERM EXIT 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- target/ns_hotplug_stress.sh@70 -- # nvmftestfini 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@488 -- # nvmfcleanup 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@117 -- # sync 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@120 -- # set +e 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@121 -- # for i in {1..20} 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:09:39.238 rmmod nvme_tcp 00:09:39.238 rmmod nvme_fabrics 00:09:39.238 rmmod nvme_keyring 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@124 -- # set -e 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@125 -- # return 0 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@489 -- # '[' -n 477568 ']' 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@490 -- # killprocess 477568 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@948 -- # '[' -z 477568 ']' 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@952 -- # kill -0 477568 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@953 -- # uname 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 477568 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@966 -- # echo 'killing process with pid 477568' 00:09:39.238 killing process with pid 477568 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@967 -- # kill 477568 00:09:39.238 17:55:32 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@972 -- # wait 477568 00:09:39.498 17:55:33 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:09:39.498 17:55:33 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:09:39.498 17:55:33 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:09:39.498 17:55:33 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:09:39.498 17:55:33 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@278 -- # remove_spdk_ns 00:09:39.498 17:55:33 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:09:39.498 17:55:33 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:09:39.498 17:55:33 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:09:41.403 17:55:35 nvmf_tcp.nvmf_ns_hotplug_stress -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:09:41.403 00:09:41.403 real 0m46.756s 00:09:41.403 user 3m18.848s 00:09:41.403 sys 0m16.622s 00:09:41.403 17:55:35 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@1124 -- # xtrace_disable 00:09:41.403 17:55:35 nvmf_tcp.nvmf_ns_hotplug_stress -- common/autotest_common.sh@10 -- # set +x 00:09:41.403 ************************************ 00:09:41.403 END TEST nvmf_ns_hotplug_stress 00:09:41.403 ************************************ 00:09:41.403 17:55:35 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:09:41.403 17:55:35 nvmf_tcp -- nvmf/nvmf.sh@33 -- # run_test nvmf_connect_stress /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_stress.sh --transport=tcp 00:09:41.403 17:55:35 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:09:41.403 17:55:35 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:09:41.403 17:55:35 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:09:41.663 ************************************ 00:09:41.663 START TEST nvmf_connect_stress 00:09:41.663 ************************************ 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_stress.sh --transport=tcp 00:09:41.663 * Looking for test storage... 00:09:41.663 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@7 -- # uname -s 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- paths/export.sh@5 -- # export PATH 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@47 -- # : 0 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@51 -- # have_pci_nics=0 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@12 -- # nvmftestinit 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@448 -- # prepare_net_devs 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@410 -- # local -g is_hw=no 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@412 -- # remove_spdk_ns 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@285 -- # xtrace_disable 00:09:41.663 17:55:35 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@291 -- # pci_devs=() 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@291 -- # local -a pci_devs 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@292 -- # pci_net_devs=() 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@293 -- # pci_drivers=() 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@293 -- # local -A pci_drivers 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@295 -- # net_devs=() 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@295 -- # local -ga net_devs 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@296 -- # e810=() 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@296 -- # local -ga e810 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@297 -- # x722=() 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@297 -- # local -ga x722 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@298 -- # mlx=() 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@298 -- # local -ga mlx 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:09:46.933 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:09:46.934 Found 0000:86:00.0 (0x8086 - 0x159b) 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:09:46.934 Found 0000:86:00.1 (0x8086 - 0x159b) 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@390 -- # [[ up == up ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:09:46.934 Found net devices under 0000:86:00.0: cvl_0_0 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@390 -- # [[ up == up ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:09:46.934 Found net devices under 0000:86:00.1: cvl_0_1 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@414 -- # is_hw=yes 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:09:46.934 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:09:46.934 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.170 ms 00:09:46.934 00:09:46.934 --- 10.0.0.2 ping statistics --- 00:09:46.934 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:46.934 rtt min/avg/max/mdev = 0.170/0.170/0.170/0.000 ms 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:09:46.934 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:09:46.934 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.215 ms 00:09:46.934 00:09:46.934 --- 10.0.0.1 ping statistics --- 00:09:46.934 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:09:46.934 rtt min/avg/max/mdev = 0.215/0.215/0.215/0.000 ms 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@422 -- # return 0 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@13 -- # nvmfappstart -m 0xE 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@722 -- # xtrace_disable 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@481 -- # nvmfpid=487735 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@482 -- # waitforlisten 487735 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@829 -- # '[' -z 487735 ']' 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@834 -- # local max_retries=100 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:09:46.934 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@838 -- # xtrace_disable 00:09:46.934 17:55:40 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:46.934 [2024-07-15 17:55:40.487186] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:09:46.934 [2024-07-15 17:55:40.487233] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:09:46.934 EAL: No free 2048 kB hugepages reported on node 1 00:09:46.934 [2024-07-15 17:55:40.545599] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:09:46.934 [2024-07-15 17:55:40.618933] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:09:46.934 [2024-07-15 17:55:40.618977] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:09:46.934 [2024-07-15 17:55:40.618984] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:09:46.934 [2024-07-15 17:55:40.618990] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:09:46.934 [2024-07-15 17:55:40.618996] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:09:46.934 [2024-07-15 17:55:40.619095] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:09:46.934 [2024-07-15 17:55:40.619183] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:09:46.934 [2024-07-15 17:55:40.619185] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@862 -- # return 0 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@728 -- # xtrace_disable 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:47.870 [2024-07-15 17:55:41.339539] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:47.870 [2024-07-15 17:55:41.362576] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:47.870 NULL1 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@21 -- # PERF_PID=487916 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@23 -- # rpcs=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.txt 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@20 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/connect_stress/connect_stress -c 0x1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' -t 10 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@25 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.txt 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # seq 1 20 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:09:47.870 EAL: No free 2048 kB hugepages reported on node 1 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@27 -- # for i in $(seq 1 20) 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@28 -- # cat 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:47.870 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:48.128 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:48.129 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:48.129 17:55:41 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:48.129 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:48.129 17:55:41 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:48.387 17:55:42 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:48.387 17:55:42 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:48.387 17:55:42 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:48.387 17:55:42 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:48.387 17:55:42 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:48.953 17:55:42 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:48.953 17:55:42 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:48.953 17:55:42 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:48.954 17:55:42 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:48.954 17:55:42 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:49.211 17:55:42 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:49.211 17:55:42 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:49.211 17:55:42 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:49.211 17:55:42 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:49.211 17:55:42 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:49.468 17:55:43 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:49.468 17:55:43 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:49.468 17:55:43 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:49.468 17:55:43 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:49.468 17:55:43 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:49.726 17:55:43 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:49.726 17:55:43 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:49.726 17:55:43 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:49.726 17:55:43 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:49.726 17:55:43 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:50.292 17:55:43 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:50.292 17:55:43 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:50.292 17:55:43 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:50.292 17:55:43 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:50.292 17:55:43 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:50.551 17:55:44 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:50.551 17:55:44 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:50.551 17:55:44 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:50.551 17:55:44 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:50.551 17:55:44 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:50.810 17:55:44 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:50.810 17:55:44 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:50.810 17:55:44 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:50.810 17:55:44 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:50.810 17:55:44 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:51.069 17:55:44 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:51.069 17:55:44 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:51.069 17:55:44 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:51.069 17:55:44 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:51.069 17:55:44 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:51.328 17:55:45 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:51.328 17:55:45 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:51.328 17:55:45 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:51.328 17:55:45 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:51.328 17:55:45 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:51.896 17:55:45 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:51.896 17:55:45 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:51.896 17:55:45 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:51.897 17:55:45 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:51.897 17:55:45 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:52.156 17:55:45 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:52.156 17:55:45 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:52.156 17:55:45 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:52.156 17:55:45 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:52.156 17:55:45 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:52.416 17:55:45 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:52.416 17:55:45 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:52.416 17:55:45 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:52.416 17:55:45 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:52.416 17:55:45 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:52.675 17:55:46 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:52.675 17:55:46 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:52.675 17:55:46 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:52.675 17:55:46 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:52.675 17:55:46 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:52.934 17:55:46 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:52.934 17:55:46 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:52.934 17:55:46 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:52.934 17:55:46 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:52.934 17:55:46 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:53.502 17:55:46 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:53.502 17:55:46 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:53.502 17:55:46 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:53.502 17:55:46 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:53.502 17:55:46 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:53.762 17:55:47 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:53.762 17:55:47 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:53.762 17:55:47 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:53.762 17:55:47 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:53.762 17:55:47 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:54.021 17:55:47 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:54.021 17:55:47 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:54.021 17:55:47 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:54.021 17:55:47 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:54.021 17:55:47 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:54.279 17:55:47 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:54.279 17:55:47 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:54.279 17:55:47 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:54.279 17:55:47 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:54.279 17:55:47 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:54.539 17:55:48 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:54.539 17:55:48 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:54.539 17:55:48 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:54.539 17:55:48 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:54.539 17:55:48 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:55.108 17:55:48 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:55.108 17:55:48 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:55.108 17:55:48 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:55.108 17:55:48 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:55.108 17:55:48 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:55.366 17:55:48 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:55.366 17:55:48 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:55.366 17:55:48 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:55.366 17:55:48 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:55.366 17:55:48 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:55.625 17:55:49 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:55.625 17:55:49 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:55.625 17:55:49 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:55.625 17:55:49 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:55.625 17:55:49 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:55.883 17:55:49 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:55.883 17:55:49 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:55.883 17:55:49 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:55.883 17:55:49 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:55.883 17:55:49 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:56.141 17:55:49 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:56.141 17:55:49 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:56.141 17:55:49 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:56.141 17:55:49 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:56.141 17:55:49 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:56.743 17:55:50 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:56.743 17:55:50 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:56.743 17:55:50 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:56.743 17:55:50 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:56.743 17:55:50 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:57.001 17:55:50 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:57.001 17:55:50 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:57.001 17:55:50 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:57.001 17:55:50 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:57.001 17:55:50 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:57.260 17:55:50 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:57.260 17:55:50 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:57.260 17:55:50 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:57.260 17:55:50 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:57.260 17:55:50 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:57.517 17:55:51 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:57.517 17:55:51 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:57.517 17:55:51 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:57.517 17:55:51 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:57.517 17:55:51 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:57.775 17:55:51 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:57.775 17:55:51 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:57.775 17:55:51 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@35 -- # rpc_cmd 00:09:57.775 17:55:51 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@559 -- # xtrace_disable 00:09:57.775 17:55:51 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:09:58.033 Testing NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@34 -- # kill -0 487916 00:09:58.291 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/connect_stress.sh: line 34: kill: (487916) - No such process 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@38 -- # wait 487916 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@39 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpc.txt 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@41 -- # trap - SIGINT SIGTERM EXIT 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- target/connect_stress.sh@43 -- # nvmftestfini 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@488 -- # nvmfcleanup 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@117 -- # sync 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@120 -- # set +e 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@121 -- # for i in {1..20} 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:09:58.291 rmmod nvme_tcp 00:09:58.291 rmmod nvme_fabrics 00:09:58.291 rmmod nvme_keyring 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@124 -- # set -e 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@125 -- # return 0 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@489 -- # '[' -n 487735 ']' 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@490 -- # killprocess 487735 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@948 -- # '[' -z 487735 ']' 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@952 -- # kill -0 487735 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@953 -- # uname 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 487735 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@966 -- # echo 'killing process with pid 487735' 00:09:58.291 killing process with pid 487735 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@967 -- # kill 487735 00:09:58.291 17:55:51 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@972 -- # wait 487735 00:09:58.550 17:55:52 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:09:58.550 17:55:52 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:09:58.550 17:55:52 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:09:58.550 17:55:52 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:09:58.550 17:55:52 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@278 -- # remove_spdk_ns 00:09:58.550 17:55:52 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:09:58.550 17:55:52 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:09:58.550 17:55:52 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:00.457 17:55:54 nvmf_tcp.nvmf_connect_stress -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:10:00.457 00:10:00.457 real 0m18.994s 00:10:00.457 user 0m41.773s 00:10:00.457 sys 0m7.768s 00:10:00.457 17:55:54 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@1124 -- # xtrace_disable 00:10:00.457 17:55:54 nvmf_tcp.nvmf_connect_stress -- common/autotest_common.sh@10 -- # set +x 00:10:00.457 ************************************ 00:10:00.457 END TEST nvmf_connect_stress 00:10:00.457 ************************************ 00:10:00.457 17:55:54 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:10:00.457 17:55:54 nvmf_tcp -- nvmf/nvmf.sh@34 -- # run_test nvmf_fused_ordering /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fused_ordering.sh --transport=tcp 00:10:00.457 17:55:54 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:10:00.457 17:55:54 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:10:00.457 17:55:54 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:10:00.717 ************************************ 00:10:00.717 START TEST nvmf_fused_ordering 00:10:00.717 ************************************ 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fused_ordering.sh --transport=tcp 00:10:00.717 * Looking for test storage... 00:10:00.717 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- target/fused_ordering.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@7 -- # uname -s 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:00.717 17:55:54 nvmf_tcp.nvmf_fused_ordering -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- paths/export.sh@5 -- # export PATH 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@47 -- # : 0 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@51 -- # have_pci_nics=0 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- target/fused_ordering.sh@12 -- # nvmftestinit 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@448 -- # prepare_net_devs 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@410 -- # local -g is_hw=no 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@412 -- # remove_spdk_ns 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@285 -- # xtrace_disable 00:10:00.718 17:55:54 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@291 -- # pci_devs=() 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@291 -- # local -a pci_devs 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@292 -- # pci_net_devs=() 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@293 -- # pci_drivers=() 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@293 -- # local -A pci_drivers 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@295 -- # net_devs=() 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@295 -- # local -ga net_devs 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@296 -- # e810=() 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@296 -- # local -ga e810 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@297 -- # x722=() 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@297 -- # local -ga x722 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@298 -- # mlx=() 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@298 -- # local -ga mlx 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:10:05.987 Found 0000:86:00.0 (0x8086 - 0x159b) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:10:05.987 Found 0000:86:00.1 (0x8086 - 0x159b) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@390 -- # [[ up == up ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:10:05.987 Found net devices under 0000:86:00.0: cvl_0_0 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@390 -- # [[ up == up ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:10:05.987 Found net devices under 0000:86:00.1: cvl_0_1 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@414 -- # is_hw=yes 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:10:05.987 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:10:05.988 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:10:05.988 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:10:05.988 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:10:05.988 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:10:05.988 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:10:05.988 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.165 ms 00:10:05.988 00:10:05.988 --- 10.0.0.2 ping statistics --- 00:10:05.988 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:05.988 rtt min/avg/max/mdev = 0.165/0.165/0.165/0.000 ms 00:10:05.988 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:10:05.988 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:10:05.988 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.244 ms 00:10:05.988 00:10:05.988 --- 10.0.0.1 ping statistics --- 00:10:05.988 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:05.988 rtt min/avg/max/mdev = 0.244/0.244/0.244/0.000 ms 00:10:05.988 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:10:05.988 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@422 -- # return 0 00:10:05.988 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:10:05.988 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:10:05.988 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:10:05.988 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:10:05.988 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:10:05.988 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:10:05.988 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:10:06.246 17:55:59 nvmf_tcp.nvmf_fused_ordering -- target/fused_ordering.sh@13 -- # nvmfappstart -m 0x2 00:10:06.246 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:10:06.246 17:55:59 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@722 -- # xtrace_disable 00:10:06.246 17:55:59 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:10:06.246 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@481 -- # nvmfpid=493074 00:10:06.246 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:10:06.246 17:55:59 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@482 -- # waitforlisten 493074 00:10:06.246 17:55:59 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@829 -- # '[' -z 493074 ']' 00:10:06.246 17:55:59 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:06.246 17:55:59 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@834 -- # local max_retries=100 00:10:06.246 17:55:59 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:06.246 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:06.246 17:55:59 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@838 -- # xtrace_disable 00:10:06.246 17:55:59 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:10:06.246 [2024-07-15 17:55:59.770446] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:10:06.246 [2024-07-15 17:55:59.770486] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:06.247 EAL: No free 2048 kB hugepages reported on node 1 00:10:06.247 [2024-07-15 17:55:59.830272] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:06.247 [2024-07-15 17:55:59.908238] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:10:06.247 [2024-07-15 17:55:59.908277] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:10:06.247 [2024-07-15 17:55:59.908284] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:10:06.247 [2024-07-15 17:55:59.908291] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:10:06.247 [2024-07-15 17:55:59.908296] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:10:06.247 [2024-07-15 17:55:59.908315] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@862 -- # return 0 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@728 -- # xtrace_disable 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- target/fused_ordering.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:10:07.185 [2024-07-15 17:56:00.619758] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- target/fused_ordering.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- target/fused_ordering.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:10:07.185 [2024-07-15 17:56:00.635904] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- target/fused_ordering.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:10:07.185 NULL1 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- target/fused_ordering.sh@19 -- # rpc_cmd bdev_wait_for_examine 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- target/fused_ordering.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 NULL1 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:07.185 17:56:00 nvmf_tcp.nvmf_fused_ordering -- target/fused_ordering.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/fused_ordering/fused_ordering -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:10:07.185 [2024-07-15 17:56:00.687956] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:10:07.185 [2024-07-15 17:56:00.687987] [ DPDK EAL parameters: fused_ordering --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid493306 ] 00:10:07.185 EAL: No free 2048 kB hugepages reported on node 1 00:10:07.444 Attached to nqn.2016-06.io.spdk:cnode1 00:10:07.445 Namespace ID: 1 size: 1GB 00:10:07.445 fused_ordering(0) 00:10:07.445 fused_ordering(1) 00:10:07.445 fused_ordering(2) 00:10:07.445 fused_ordering(3) 00:10:07.445 fused_ordering(4) 00:10:07.445 fused_ordering(5) 00:10:07.445 fused_ordering(6) 00:10:07.445 fused_ordering(7) 00:10:07.445 fused_ordering(8) 00:10:07.445 fused_ordering(9) 00:10:07.445 fused_ordering(10) 00:10:07.445 fused_ordering(11) 00:10:07.445 fused_ordering(12) 00:10:07.445 fused_ordering(13) 00:10:07.445 fused_ordering(14) 00:10:07.445 fused_ordering(15) 00:10:07.445 fused_ordering(16) 00:10:07.445 fused_ordering(17) 00:10:07.445 fused_ordering(18) 00:10:07.445 fused_ordering(19) 00:10:07.445 fused_ordering(20) 00:10:07.445 fused_ordering(21) 00:10:07.445 fused_ordering(22) 00:10:07.445 fused_ordering(23) 00:10:07.445 fused_ordering(24) 00:10:07.445 fused_ordering(25) 00:10:07.445 fused_ordering(26) 00:10:07.445 fused_ordering(27) 00:10:07.445 fused_ordering(28) 00:10:07.445 fused_ordering(29) 00:10:07.445 fused_ordering(30) 00:10:07.445 fused_ordering(31) 00:10:07.445 fused_ordering(32) 00:10:07.445 fused_ordering(33) 00:10:07.445 fused_ordering(34) 00:10:07.445 fused_ordering(35) 00:10:07.445 fused_ordering(36) 00:10:07.445 fused_ordering(37) 00:10:07.445 fused_ordering(38) 00:10:07.445 fused_ordering(39) 00:10:07.445 fused_ordering(40) 00:10:07.445 fused_ordering(41) 00:10:07.445 fused_ordering(42) 00:10:07.445 fused_ordering(43) 00:10:07.445 fused_ordering(44) 00:10:07.445 fused_ordering(45) 00:10:07.445 fused_ordering(46) 00:10:07.445 fused_ordering(47) 00:10:07.445 fused_ordering(48) 00:10:07.445 fused_ordering(49) 00:10:07.445 fused_ordering(50) 00:10:07.445 fused_ordering(51) 00:10:07.445 fused_ordering(52) 00:10:07.445 fused_ordering(53) 00:10:07.445 fused_ordering(54) 00:10:07.445 fused_ordering(55) 00:10:07.445 fused_ordering(56) 00:10:07.445 fused_ordering(57) 00:10:07.445 fused_ordering(58) 00:10:07.445 fused_ordering(59) 00:10:07.445 fused_ordering(60) 00:10:07.445 fused_ordering(61) 00:10:07.445 fused_ordering(62) 00:10:07.445 fused_ordering(63) 00:10:07.445 fused_ordering(64) 00:10:07.445 fused_ordering(65) 00:10:07.445 fused_ordering(66) 00:10:07.445 fused_ordering(67) 00:10:07.445 fused_ordering(68) 00:10:07.445 fused_ordering(69) 00:10:07.445 fused_ordering(70) 00:10:07.445 fused_ordering(71) 00:10:07.445 fused_ordering(72) 00:10:07.445 fused_ordering(73) 00:10:07.445 fused_ordering(74) 00:10:07.445 fused_ordering(75) 00:10:07.445 fused_ordering(76) 00:10:07.445 fused_ordering(77) 00:10:07.445 fused_ordering(78) 00:10:07.445 fused_ordering(79) 00:10:07.445 fused_ordering(80) 00:10:07.445 fused_ordering(81) 00:10:07.445 fused_ordering(82) 00:10:07.445 fused_ordering(83) 00:10:07.445 fused_ordering(84) 00:10:07.445 fused_ordering(85) 00:10:07.445 fused_ordering(86) 00:10:07.445 fused_ordering(87) 00:10:07.445 fused_ordering(88) 00:10:07.445 fused_ordering(89) 00:10:07.445 fused_ordering(90) 00:10:07.445 fused_ordering(91) 00:10:07.445 fused_ordering(92) 00:10:07.445 fused_ordering(93) 00:10:07.445 fused_ordering(94) 00:10:07.445 fused_ordering(95) 00:10:07.445 fused_ordering(96) 00:10:07.445 fused_ordering(97) 00:10:07.445 fused_ordering(98) 00:10:07.445 fused_ordering(99) 00:10:07.445 fused_ordering(100) 00:10:07.445 fused_ordering(101) 00:10:07.445 fused_ordering(102) 00:10:07.445 fused_ordering(103) 00:10:07.445 fused_ordering(104) 00:10:07.445 fused_ordering(105) 00:10:07.445 fused_ordering(106) 00:10:07.445 fused_ordering(107) 00:10:07.445 fused_ordering(108) 00:10:07.445 fused_ordering(109) 00:10:07.445 fused_ordering(110) 00:10:07.445 fused_ordering(111) 00:10:07.445 fused_ordering(112) 00:10:07.445 fused_ordering(113) 00:10:07.445 fused_ordering(114) 00:10:07.445 fused_ordering(115) 00:10:07.445 fused_ordering(116) 00:10:07.445 fused_ordering(117) 00:10:07.445 fused_ordering(118) 00:10:07.445 fused_ordering(119) 00:10:07.445 fused_ordering(120) 00:10:07.445 fused_ordering(121) 00:10:07.445 fused_ordering(122) 00:10:07.445 fused_ordering(123) 00:10:07.445 fused_ordering(124) 00:10:07.445 fused_ordering(125) 00:10:07.445 fused_ordering(126) 00:10:07.445 fused_ordering(127) 00:10:07.445 fused_ordering(128) 00:10:07.445 fused_ordering(129) 00:10:07.445 fused_ordering(130) 00:10:07.445 fused_ordering(131) 00:10:07.445 fused_ordering(132) 00:10:07.445 fused_ordering(133) 00:10:07.445 fused_ordering(134) 00:10:07.445 fused_ordering(135) 00:10:07.445 fused_ordering(136) 00:10:07.445 fused_ordering(137) 00:10:07.445 fused_ordering(138) 00:10:07.445 fused_ordering(139) 00:10:07.445 fused_ordering(140) 00:10:07.445 fused_ordering(141) 00:10:07.445 fused_ordering(142) 00:10:07.445 fused_ordering(143) 00:10:07.445 fused_ordering(144) 00:10:07.445 fused_ordering(145) 00:10:07.445 fused_ordering(146) 00:10:07.445 fused_ordering(147) 00:10:07.445 fused_ordering(148) 00:10:07.445 fused_ordering(149) 00:10:07.445 fused_ordering(150) 00:10:07.445 fused_ordering(151) 00:10:07.445 fused_ordering(152) 00:10:07.445 fused_ordering(153) 00:10:07.445 fused_ordering(154) 00:10:07.445 fused_ordering(155) 00:10:07.445 fused_ordering(156) 00:10:07.445 fused_ordering(157) 00:10:07.445 fused_ordering(158) 00:10:07.445 fused_ordering(159) 00:10:07.445 fused_ordering(160) 00:10:07.445 fused_ordering(161) 00:10:07.445 fused_ordering(162) 00:10:07.445 fused_ordering(163) 00:10:07.445 fused_ordering(164) 00:10:07.445 fused_ordering(165) 00:10:07.445 fused_ordering(166) 00:10:07.445 fused_ordering(167) 00:10:07.445 fused_ordering(168) 00:10:07.445 fused_ordering(169) 00:10:07.445 fused_ordering(170) 00:10:07.445 fused_ordering(171) 00:10:07.445 fused_ordering(172) 00:10:07.445 fused_ordering(173) 00:10:07.445 fused_ordering(174) 00:10:07.445 fused_ordering(175) 00:10:07.445 fused_ordering(176) 00:10:07.445 fused_ordering(177) 00:10:07.445 fused_ordering(178) 00:10:07.445 fused_ordering(179) 00:10:07.445 fused_ordering(180) 00:10:07.445 fused_ordering(181) 00:10:07.445 fused_ordering(182) 00:10:07.445 fused_ordering(183) 00:10:07.445 fused_ordering(184) 00:10:07.445 fused_ordering(185) 00:10:07.445 fused_ordering(186) 00:10:07.445 fused_ordering(187) 00:10:07.445 fused_ordering(188) 00:10:07.445 fused_ordering(189) 00:10:07.445 fused_ordering(190) 00:10:07.445 fused_ordering(191) 00:10:07.445 fused_ordering(192) 00:10:07.445 fused_ordering(193) 00:10:07.445 fused_ordering(194) 00:10:07.445 fused_ordering(195) 00:10:07.445 fused_ordering(196) 00:10:07.445 fused_ordering(197) 00:10:07.445 fused_ordering(198) 00:10:07.445 fused_ordering(199) 00:10:07.445 fused_ordering(200) 00:10:07.445 fused_ordering(201) 00:10:07.445 fused_ordering(202) 00:10:07.445 fused_ordering(203) 00:10:07.445 fused_ordering(204) 00:10:07.445 fused_ordering(205) 00:10:07.705 fused_ordering(206) 00:10:07.705 fused_ordering(207) 00:10:07.705 fused_ordering(208) 00:10:07.705 fused_ordering(209) 00:10:07.705 fused_ordering(210) 00:10:07.705 fused_ordering(211) 00:10:07.705 fused_ordering(212) 00:10:07.705 fused_ordering(213) 00:10:07.705 fused_ordering(214) 00:10:07.705 fused_ordering(215) 00:10:07.705 fused_ordering(216) 00:10:07.705 fused_ordering(217) 00:10:07.705 fused_ordering(218) 00:10:07.705 fused_ordering(219) 00:10:07.705 fused_ordering(220) 00:10:07.705 fused_ordering(221) 00:10:07.705 fused_ordering(222) 00:10:07.705 fused_ordering(223) 00:10:07.705 fused_ordering(224) 00:10:07.705 fused_ordering(225) 00:10:07.705 fused_ordering(226) 00:10:07.705 fused_ordering(227) 00:10:07.705 fused_ordering(228) 00:10:07.705 fused_ordering(229) 00:10:07.705 fused_ordering(230) 00:10:07.705 fused_ordering(231) 00:10:07.705 fused_ordering(232) 00:10:07.705 fused_ordering(233) 00:10:07.705 fused_ordering(234) 00:10:07.705 fused_ordering(235) 00:10:07.705 fused_ordering(236) 00:10:07.705 fused_ordering(237) 00:10:07.705 fused_ordering(238) 00:10:07.705 fused_ordering(239) 00:10:07.705 fused_ordering(240) 00:10:07.705 fused_ordering(241) 00:10:07.705 fused_ordering(242) 00:10:07.705 fused_ordering(243) 00:10:07.705 fused_ordering(244) 00:10:07.705 fused_ordering(245) 00:10:07.705 fused_ordering(246) 00:10:07.705 fused_ordering(247) 00:10:07.705 fused_ordering(248) 00:10:07.705 fused_ordering(249) 00:10:07.705 fused_ordering(250) 00:10:07.705 fused_ordering(251) 00:10:07.705 fused_ordering(252) 00:10:07.705 fused_ordering(253) 00:10:07.705 fused_ordering(254) 00:10:07.705 fused_ordering(255) 00:10:07.705 fused_ordering(256) 00:10:07.705 fused_ordering(257) 00:10:07.705 fused_ordering(258) 00:10:07.705 fused_ordering(259) 00:10:07.705 fused_ordering(260) 00:10:07.705 fused_ordering(261) 00:10:07.705 fused_ordering(262) 00:10:07.705 fused_ordering(263) 00:10:07.705 fused_ordering(264) 00:10:07.705 fused_ordering(265) 00:10:07.705 fused_ordering(266) 00:10:07.705 fused_ordering(267) 00:10:07.705 fused_ordering(268) 00:10:07.705 fused_ordering(269) 00:10:07.705 fused_ordering(270) 00:10:07.705 fused_ordering(271) 00:10:07.705 fused_ordering(272) 00:10:07.705 fused_ordering(273) 00:10:07.705 fused_ordering(274) 00:10:07.705 fused_ordering(275) 00:10:07.705 fused_ordering(276) 00:10:07.705 fused_ordering(277) 00:10:07.705 fused_ordering(278) 00:10:07.705 fused_ordering(279) 00:10:07.705 fused_ordering(280) 00:10:07.705 fused_ordering(281) 00:10:07.705 fused_ordering(282) 00:10:07.705 fused_ordering(283) 00:10:07.705 fused_ordering(284) 00:10:07.705 fused_ordering(285) 00:10:07.705 fused_ordering(286) 00:10:07.705 fused_ordering(287) 00:10:07.705 fused_ordering(288) 00:10:07.705 fused_ordering(289) 00:10:07.705 fused_ordering(290) 00:10:07.705 fused_ordering(291) 00:10:07.705 fused_ordering(292) 00:10:07.705 fused_ordering(293) 00:10:07.705 fused_ordering(294) 00:10:07.705 fused_ordering(295) 00:10:07.705 fused_ordering(296) 00:10:07.705 fused_ordering(297) 00:10:07.705 fused_ordering(298) 00:10:07.705 fused_ordering(299) 00:10:07.705 fused_ordering(300) 00:10:07.705 fused_ordering(301) 00:10:07.705 fused_ordering(302) 00:10:07.705 fused_ordering(303) 00:10:07.705 fused_ordering(304) 00:10:07.705 fused_ordering(305) 00:10:07.705 fused_ordering(306) 00:10:07.705 fused_ordering(307) 00:10:07.705 fused_ordering(308) 00:10:07.705 fused_ordering(309) 00:10:07.705 fused_ordering(310) 00:10:07.705 fused_ordering(311) 00:10:07.705 fused_ordering(312) 00:10:07.705 fused_ordering(313) 00:10:07.705 fused_ordering(314) 00:10:07.705 fused_ordering(315) 00:10:07.705 fused_ordering(316) 00:10:07.705 fused_ordering(317) 00:10:07.705 fused_ordering(318) 00:10:07.705 fused_ordering(319) 00:10:07.705 fused_ordering(320) 00:10:07.705 fused_ordering(321) 00:10:07.705 fused_ordering(322) 00:10:07.705 fused_ordering(323) 00:10:07.705 fused_ordering(324) 00:10:07.705 fused_ordering(325) 00:10:07.705 fused_ordering(326) 00:10:07.705 fused_ordering(327) 00:10:07.705 fused_ordering(328) 00:10:07.705 fused_ordering(329) 00:10:07.705 fused_ordering(330) 00:10:07.705 fused_ordering(331) 00:10:07.705 fused_ordering(332) 00:10:07.705 fused_ordering(333) 00:10:07.705 fused_ordering(334) 00:10:07.705 fused_ordering(335) 00:10:07.705 fused_ordering(336) 00:10:07.705 fused_ordering(337) 00:10:07.705 fused_ordering(338) 00:10:07.705 fused_ordering(339) 00:10:07.705 fused_ordering(340) 00:10:07.705 fused_ordering(341) 00:10:07.705 fused_ordering(342) 00:10:07.705 fused_ordering(343) 00:10:07.705 fused_ordering(344) 00:10:07.705 fused_ordering(345) 00:10:07.705 fused_ordering(346) 00:10:07.705 fused_ordering(347) 00:10:07.705 fused_ordering(348) 00:10:07.705 fused_ordering(349) 00:10:07.705 fused_ordering(350) 00:10:07.705 fused_ordering(351) 00:10:07.705 fused_ordering(352) 00:10:07.705 fused_ordering(353) 00:10:07.705 fused_ordering(354) 00:10:07.705 fused_ordering(355) 00:10:07.705 fused_ordering(356) 00:10:07.705 fused_ordering(357) 00:10:07.705 fused_ordering(358) 00:10:07.705 fused_ordering(359) 00:10:07.705 fused_ordering(360) 00:10:07.705 fused_ordering(361) 00:10:07.705 fused_ordering(362) 00:10:07.705 fused_ordering(363) 00:10:07.705 fused_ordering(364) 00:10:07.705 fused_ordering(365) 00:10:07.705 fused_ordering(366) 00:10:07.705 fused_ordering(367) 00:10:07.705 fused_ordering(368) 00:10:07.706 fused_ordering(369) 00:10:07.706 fused_ordering(370) 00:10:07.706 fused_ordering(371) 00:10:07.706 fused_ordering(372) 00:10:07.706 fused_ordering(373) 00:10:07.706 fused_ordering(374) 00:10:07.706 fused_ordering(375) 00:10:07.706 fused_ordering(376) 00:10:07.706 fused_ordering(377) 00:10:07.706 fused_ordering(378) 00:10:07.706 fused_ordering(379) 00:10:07.706 fused_ordering(380) 00:10:07.706 fused_ordering(381) 00:10:07.706 fused_ordering(382) 00:10:07.706 fused_ordering(383) 00:10:07.706 fused_ordering(384) 00:10:07.706 fused_ordering(385) 00:10:07.706 fused_ordering(386) 00:10:07.706 fused_ordering(387) 00:10:07.706 fused_ordering(388) 00:10:07.706 fused_ordering(389) 00:10:07.706 fused_ordering(390) 00:10:07.706 fused_ordering(391) 00:10:07.706 fused_ordering(392) 00:10:07.706 fused_ordering(393) 00:10:07.706 fused_ordering(394) 00:10:07.706 fused_ordering(395) 00:10:07.706 fused_ordering(396) 00:10:07.706 fused_ordering(397) 00:10:07.706 fused_ordering(398) 00:10:07.706 fused_ordering(399) 00:10:07.706 fused_ordering(400) 00:10:07.706 fused_ordering(401) 00:10:07.706 fused_ordering(402) 00:10:07.706 fused_ordering(403) 00:10:07.706 fused_ordering(404) 00:10:07.706 fused_ordering(405) 00:10:07.706 fused_ordering(406) 00:10:07.706 fused_ordering(407) 00:10:07.706 fused_ordering(408) 00:10:07.706 fused_ordering(409) 00:10:07.706 fused_ordering(410) 00:10:07.965 fused_ordering(411) 00:10:07.965 fused_ordering(412) 00:10:07.965 fused_ordering(413) 00:10:07.965 fused_ordering(414) 00:10:07.965 fused_ordering(415) 00:10:07.965 fused_ordering(416) 00:10:07.965 fused_ordering(417) 00:10:07.965 fused_ordering(418) 00:10:07.965 fused_ordering(419) 00:10:07.965 fused_ordering(420) 00:10:07.965 fused_ordering(421) 00:10:07.965 fused_ordering(422) 00:10:07.965 fused_ordering(423) 00:10:07.965 fused_ordering(424) 00:10:07.965 fused_ordering(425) 00:10:07.965 fused_ordering(426) 00:10:07.965 fused_ordering(427) 00:10:07.965 fused_ordering(428) 00:10:07.965 fused_ordering(429) 00:10:07.965 fused_ordering(430) 00:10:07.965 fused_ordering(431) 00:10:07.965 fused_ordering(432) 00:10:07.965 fused_ordering(433) 00:10:07.965 fused_ordering(434) 00:10:07.965 fused_ordering(435) 00:10:07.965 fused_ordering(436) 00:10:07.965 fused_ordering(437) 00:10:07.965 fused_ordering(438) 00:10:07.965 fused_ordering(439) 00:10:07.965 fused_ordering(440) 00:10:07.965 fused_ordering(441) 00:10:07.965 fused_ordering(442) 00:10:07.965 fused_ordering(443) 00:10:07.965 fused_ordering(444) 00:10:07.965 fused_ordering(445) 00:10:07.965 fused_ordering(446) 00:10:07.965 fused_ordering(447) 00:10:07.965 fused_ordering(448) 00:10:07.965 fused_ordering(449) 00:10:07.965 fused_ordering(450) 00:10:07.965 fused_ordering(451) 00:10:07.965 fused_ordering(452) 00:10:07.965 fused_ordering(453) 00:10:07.965 fused_ordering(454) 00:10:07.965 fused_ordering(455) 00:10:07.965 fused_ordering(456) 00:10:07.965 fused_ordering(457) 00:10:07.965 fused_ordering(458) 00:10:07.965 fused_ordering(459) 00:10:07.965 fused_ordering(460) 00:10:07.965 fused_ordering(461) 00:10:07.965 fused_ordering(462) 00:10:07.965 fused_ordering(463) 00:10:07.965 fused_ordering(464) 00:10:07.965 fused_ordering(465) 00:10:07.965 fused_ordering(466) 00:10:07.965 fused_ordering(467) 00:10:07.965 fused_ordering(468) 00:10:07.965 fused_ordering(469) 00:10:07.965 fused_ordering(470) 00:10:07.965 fused_ordering(471) 00:10:07.965 fused_ordering(472) 00:10:07.965 fused_ordering(473) 00:10:07.965 fused_ordering(474) 00:10:07.965 fused_ordering(475) 00:10:07.965 fused_ordering(476) 00:10:07.965 fused_ordering(477) 00:10:07.965 fused_ordering(478) 00:10:07.965 fused_ordering(479) 00:10:07.965 fused_ordering(480) 00:10:07.965 fused_ordering(481) 00:10:07.965 fused_ordering(482) 00:10:07.965 fused_ordering(483) 00:10:07.965 fused_ordering(484) 00:10:07.965 fused_ordering(485) 00:10:07.965 fused_ordering(486) 00:10:07.965 fused_ordering(487) 00:10:07.965 fused_ordering(488) 00:10:07.965 fused_ordering(489) 00:10:07.965 fused_ordering(490) 00:10:07.965 fused_ordering(491) 00:10:07.965 fused_ordering(492) 00:10:07.965 fused_ordering(493) 00:10:07.965 fused_ordering(494) 00:10:07.965 fused_ordering(495) 00:10:07.965 fused_ordering(496) 00:10:07.965 fused_ordering(497) 00:10:07.965 fused_ordering(498) 00:10:07.965 fused_ordering(499) 00:10:07.965 fused_ordering(500) 00:10:07.965 fused_ordering(501) 00:10:07.965 fused_ordering(502) 00:10:07.965 fused_ordering(503) 00:10:07.965 fused_ordering(504) 00:10:07.965 fused_ordering(505) 00:10:07.965 fused_ordering(506) 00:10:07.965 fused_ordering(507) 00:10:07.965 fused_ordering(508) 00:10:07.965 fused_ordering(509) 00:10:07.965 fused_ordering(510) 00:10:07.965 fused_ordering(511) 00:10:07.965 fused_ordering(512) 00:10:07.965 fused_ordering(513) 00:10:07.965 fused_ordering(514) 00:10:07.965 fused_ordering(515) 00:10:07.965 fused_ordering(516) 00:10:07.965 fused_ordering(517) 00:10:07.965 fused_ordering(518) 00:10:07.965 fused_ordering(519) 00:10:07.965 fused_ordering(520) 00:10:07.965 fused_ordering(521) 00:10:07.965 fused_ordering(522) 00:10:07.965 fused_ordering(523) 00:10:07.965 fused_ordering(524) 00:10:07.965 fused_ordering(525) 00:10:07.965 fused_ordering(526) 00:10:07.965 fused_ordering(527) 00:10:07.965 fused_ordering(528) 00:10:07.965 fused_ordering(529) 00:10:07.965 fused_ordering(530) 00:10:07.965 fused_ordering(531) 00:10:07.965 fused_ordering(532) 00:10:07.965 fused_ordering(533) 00:10:07.965 fused_ordering(534) 00:10:07.965 fused_ordering(535) 00:10:07.965 fused_ordering(536) 00:10:07.965 fused_ordering(537) 00:10:07.965 fused_ordering(538) 00:10:07.965 fused_ordering(539) 00:10:07.965 fused_ordering(540) 00:10:07.965 fused_ordering(541) 00:10:07.965 fused_ordering(542) 00:10:07.965 fused_ordering(543) 00:10:07.965 fused_ordering(544) 00:10:07.965 fused_ordering(545) 00:10:07.965 fused_ordering(546) 00:10:07.965 fused_ordering(547) 00:10:07.965 fused_ordering(548) 00:10:07.965 fused_ordering(549) 00:10:07.965 fused_ordering(550) 00:10:07.965 fused_ordering(551) 00:10:07.965 fused_ordering(552) 00:10:07.965 fused_ordering(553) 00:10:07.965 fused_ordering(554) 00:10:07.965 fused_ordering(555) 00:10:07.965 fused_ordering(556) 00:10:07.965 fused_ordering(557) 00:10:07.965 fused_ordering(558) 00:10:07.965 fused_ordering(559) 00:10:07.965 fused_ordering(560) 00:10:07.965 fused_ordering(561) 00:10:07.965 fused_ordering(562) 00:10:07.965 fused_ordering(563) 00:10:07.965 fused_ordering(564) 00:10:07.965 fused_ordering(565) 00:10:07.965 fused_ordering(566) 00:10:07.965 fused_ordering(567) 00:10:07.965 fused_ordering(568) 00:10:07.965 fused_ordering(569) 00:10:07.965 fused_ordering(570) 00:10:07.965 fused_ordering(571) 00:10:07.965 fused_ordering(572) 00:10:07.965 fused_ordering(573) 00:10:07.965 fused_ordering(574) 00:10:07.965 fused_ordering(575) 00:10:07.965 fused_ordering(576) 00:10:07.965 fused_ordering(577) 00:10:07.965 fused_ordering(578) 00:10:07.965 fused_ordering(579) 00:10:07.965 fused_ordering(580) 00:10:07.965 fused_ordering(581) 00:10:07.965 fused_ordering(582) 00:10:07.965 fused_ordering(583) 00:10:07.965 fused_ordering(584) 00:10:07.965 fused_ordering(585) 00:10:07.965 fused_ordering(586) 00:10:07.965 fused_ordering(587) 00:10:07.965 fused_ordering(588) 00:10:07.965 fused_ordering(589) 00:10:07.965 fused_ordering(590) 00:10:07.965 fused_ordering(591) 00:10:07.965 fused_ordering(592) 00:10:07.965 fused_ordering(593) 00:10:07.965 fused_ordering(594) 00:10:07.965 fused_ordering(595) 00:10:07.965 fused_ordering(596) 00:10:07.965 fused_ordering(597) 00:10:07.965 fused_ordering(598) 00:10:07.965 fused_ordering(599) 00:10:07.965 fused_ordering(600) 00:10:07.965 fused_ordering(601) 00:10:07.965 fused_ordering(602) 00:10:07.965 fused_ordering(603) 00:10:07.965 fused_ordering(604) 00:10:07.965 fused_ordering(605) 00:10:07.965 fused_ordering(606) 00:10:07.965 fused_ordering(607) 00:10:07.965 fused_ordering(608) 00:10:07.965 fused_ordering(609) 00:10:07.965 fused_ordering(610) 00:10:07.965 fused_ordering(611) 00:10:07.965 fused_ordering(612) 00:10:07.965 fused_ordering(613) 00:10:07.965 fused_ordering(614) 00:10:07.965 fused_ordering(615) 00:10:08.533 fused_ordering(616) 00:10:08.533 fused_ordering(617) 00:10:08.533 fused_ordering(618) 00:10:08.533 fused_ordering(619) 00:10:08.533 fused_ordering(620) 00:10:08.533 fused_ordering(621) 00:10:08.533 fused_ordering(622) 00:10:08.533 fused_ordering(623) 00:10:08.533 fused_ordering(624) 00:10:08.533 fused_ordering(625) 00:10:08.533 fused_ordering(626) 00:10:08.533 fused_ordering(627) 00:10:08.533 fused_ordering(628) 00:10:08.533 fused_ordering(629) 00:10:08.533 fused_ordering(630) 00:10:08.533 fused_ordering(631) 00:10:08.533 fused_ordering(632) 00:10:08.533 fused_ordering(633) 00:10:08.533 fused_ordering(634) 00:10:08.533 fused_ordering(635) 00:10:08.533 fused_ordering(636) 00:10:08.533 fused_ordering(637) 00:10:08.533 fused_ordering(638) 00:10:08.533 fused_ordering(639) 00:10:08.533 fused_ordering(640) 00:10:08.533 fused_ordering(641) 00:10:08.533 fused_ordering(642) 00:10:08.533 fused_ordering(643) 00:10:08.533 fused_ordering(644) 00:10:08.533 fused_ordering(645) 00:10:08.533 fused_ordering(646) 00:10:08.533 fused_ordering(647) 00:10:08.533 fused_ordering(648) 00:10:08.533 fused_ordering(649) 00:10:08.533 fused_ordering(650) 00:10:08.533 fused_ordering(651) 00:10:08.533 fused_ordering(652) 00:10:08.533 fused_ordering(653) 00:10:08.533 fused_ordering(654) 00:10:08.533 fused_ordering(655) 00:10:08.533 fused_ordering(656) 00:10:08.533 fused_ordering(657) 00:10:08.533 fused_ordering(658) 00:10:08.533 fused_ordering(659) 00:10:08.533 fused_ordering(660) 00:10:08.533 fused_ordering(661) 00:10:08.533 fused_ordering(662) 00:10:08.533 fused_ordering(663) 00:10:08.533 fused_ordering(664) 00:10:08.533 fused_ordering(665) 00:10:08.533 fused_ordering(666) 00:10:08.533 fused_ordering(667) 00:10:08.533 fused_ordering(668) 00:10:08.533 fused_ordering(669) 00:10:08.533 fused_ordering(670) 00:10:08.533 fused_ordering(671) 00:10:08.533 fused_ordering(672) 00:10:08.533 fused_ordering(673) 00:10:08.533 fused_ordering(674) 00:10:08.533 fused_ordering(675) 00:10:08.533 fused_ordering(676) 00:10:08.533 fused_ordering(677) 00:10:08.533 fused_ordering(678) 00:10:08.533 fused_ordering(679) 00:10:08.533 fused_ordering(680) 00:10:08.533 fused_ordering(681) 00:10:08.533 fused_ordering(682) 00:10:08.533 fused_ordering(683) 00:10:08.533 fused_ordering(684) 00:10:08.533 fused_ordering(685) 00:10:08.533 fused_ordering(686) 00:10:08.533 fused_ordering(687) 00:10:08.533 fused_ordering(688) 00:10:08.533 fused_ordering(689) 00:10:08.533 fused_ordering(690) 00:10:08.533 fused_ordering(691) 00:10:08.533 fused_ordering(692) 00:10:08.533 fused_ordering(693) 00:10:08.533 fused_ordering(694) 00:10:08.533 fused_ordering(695) 00:10:08.533 fused_ordering(696) 00:10:08.533 fused_ordering(697) 00:10:08.533 fused_ordering(698) 00:10:08.533 fused_ordering(699) 00:10:08.533 fused_ordering(700) 00:10:08.533 fused_ordering(701) 00:10:08.533 fused_ordering(702) 00:10:08.533 fused_ordering(703) 00:10:08.533 fused_ordering(704) 00:10:08.533 fused_ordering(705) 00:10:08.533 fused_ordering(706) 00:10:08.533 fused_ordering(707) 00:10:08.533 fused_ordering(708) 00:10:08.533 fused_ordering(709) 00:10:08.533 fused_ordering(710) 00:10:08.533 fused_ordering(711) 00:10:08.533 fused_ordering(712) 00:10:08.533 fused_ordering(713) 00:10:08.533 fused_ordering(714) 00:10:08.533 fused_ordering(715) 00:10:08.533 fused_ordering(716) 00:10:08.533 fused_ordering(717) 00:10:08.533 fused_ordering(718) 00:10:08.533 fused_ordering(719) 00:10:08.533 fused_ordering(720) 00:10:08.533 fused_ordering(721) 00:10:08.533 fused_ordering(722) 00:10:08.533 fused_ordering(723) 00:10:08.533 fused_ordering(724) 00:10:08.533 fused_ordering(725) 00:10:08.533 fused_ordering(726) 00:10:08.533 fused_ordering(727) 00:10:08.533 fused_ordering(728) 00:10:08.533 fused_ordering(729) 00:10:08.533 fused_ordering(730) 00:10:08.533 fused_ordering(731) 00:10:08.533 fused_ordering(732) 00:10:08.533 fused_ordering(733) 00:10:08.533 fused_ordering(734) 00:10:08.533 fused_ordering(735) 00:10:08.533 fused_ordering(736) 00:10:08.533 fused_ordering(737) 00:10:08.533 fused_ordering(738) 00:10:08.533 fused_ordering(739) 00:10:08.533 fused_ordering(740) 00:10:08.533 fused_ordering(741) 00:10:08.533 fused_ordering(742) 00:10:08.533 fused_ordering(743) 00:10:08.533 fused_ordering(744) 00:10:08.533 fused_ordering(745) 00:10:08.533 fused_ordering(746) 00:10:08.533 fused_ordering(747) 00:10:08.533 fused_ordering(748) 00:10:08.533 fused_ordering(749) 00:10:08.533 fused_ordering(750) 00:10:08.533 fused_ordering(751) 00:10:08.533 fused_ordering(752) 00:10:08.533 fused_ordering(753) 00:10:08.533 fused_ordering(754) 00:10:08.533 fused_ordering(755) 00:10:08.533 fused_ordering(756) 00:10:08.533 fused_ordering(757) 00:10:08.533 fused_ordering(758) 00:10:08.533 fused_ordering(759) 00:10:08.533 fused_ordering(760) 00:10:08.533 fused_ordering(761) 00:10:08.533 fused_ordering(762) 00:10:08.533 fused_ordering(763) 00:10:08.533 fused_ordering(764) 00:10:08.533 fused_ordering(765) 00:10:08.533 fused_ordering(766) 00:10:08.533 fused_ordering(767) 00:10:08.533 fused_ordering(768) 00:10:08.533 fused_ordering(769) 00:10:08.533 fused_ordering(770) 00:10:08.533 fused_ordering(771) 00:10:08.533 fused_ordering(772) 00:10:08.533 fused_ordering(773) 00:10:08.533 fused_ordering(774) 00:10:08.534 fused_ordering(775) 00:10:08.534 fused_ordering(776) 00:10:08.534 fused_ordering(777) 00:10:08.534 fused_ordering(778) 00:10:08.534 fused_ordering(779) 00:10:08.534 fused_ordering(780) 00:10:08.534 fused_ordering(781) 00:10:08.534 fused_ordering(782) 00:10:08.534 fused_ordering(783) 00:10:08.534 fused_ordering(784) 00:10:08.534 fused_ordering(785) 00:10:08.534 fused_ordering(786) 00:10:08.534 fused_ordering(787) 00:10:08.534 fused_ordering(788) 00:10:08.534 fused_ordering(789) 00:10:08.534 fused_ordering(790) 00:10:08.534 fused_ordering(791) 00:10:08.534 fused_ordering(792) 00:10:08.534 fused_ordering(793) 00:10:08.534 fused_ordering(794) 00:10:08.534 fused_ordering(795) 00:10:08.534 fused_ordering(796) 00:10:08.534 fused_ordering(797) 00:10:08.534 fused_ordering(798) 00:10:08.534 fused_ordering(799) 00:10:08.534 fused_ordering(800) 00:10:08.534 fused_ordering(801) 00:10:08.534 fused_ordering(802) 00:10:08.534 fused_ordering(803) 00:10:08.534 fused_ordering(804) 00:10:08.534 fused_ordering(805) 00:10:08.534 fused_ordering(806) 00:10:08.534 fused_ordering(807) 00:10:08.534 fused_ordering(808) 00:10:08.534 fused_ordering(809) 00:10:08.534 fused_ordering(810) 00:10:08.534 fused_ordering(811) 00:10:08.534 fused_ordering(812) 00:10:08.534 fused_ordering(813) 00:10:08.534 fused_ordering(814) 00:10:08.534 fused_ordering(815) 00:10:08.534 fused_ordering(816) 00:10:08.534 fused_ordering(817) 00:10:08.534 fused_ordering(818) 00:10:08.534 fused_ordering(819) 00:10:08.534 fused_ordering(820) 00:10:09.100 fused_ordering(821) 00:10:09.100 fused_ordering(822) 00:10:09.100 fused_ordering(823) 00:10:09.100 fused_ordering(824) 00:10:09.100 fused_ordering(825) 00:10:09.100 fused_ordering(826) 00:10:09.100 fused_ordering(827) 00:10:09.100 fused_ordering(828) 00:10:09.100 fused_ordering(829) 00:10:09.100 fused_ordering(830) 00:10:09.100 fused_ordering(831) 00:10:09.100 fused_ordering(832) 00:10:09.100 fused_ordering(833) 00:10:09.100 fused_ordering(834) 00:10:09.100 fused_ordering(835) 00:10:09.100 fused_ordering(836) 00:10:09.100 fused_ordering(837) 00:10:09.100 fused_ordering(838) 00:10:09.100 fused_ordering(839) 00:10:09.100 fused_ordering(840) 00:10:09.100 fused_ordering(841) 00:10:09.100 fused_ordering(842) 00:10:09.100 fused_ordering(843) 00:10:09.100 fused_ordering(844) 00:10:09.100 fused_ordering(845) 00:10:09.100 fused_ordering(846) 00:10:09.100 fused_ordering(847) 00:10:09.100 fused_ordering(848) 00:10:09.100 fused_ordering(849) 00:10:09.100 fused_ordering(850) 00:10:09.100 fused_ordering(851) 00:10:09.100 fused_ordering(852) 00:10:09.100 fused_ordering(853) 00:10:09.100 fused_ordering(854) 00:10:09.100 fused_ordering(855) 00:10:09.100 fused_ordering(856) 00:10:09.100 fused_ordering(857) 00:10:09.100 fused_ordering(858) 00:10:09.100 fused_ordering(859) 00:10:09.100 fused_ordering(860) 00:10:09.101 fused_ordering(861) 00:10:09.101 fused_ordering(862) 00:10:09.101 fused_ordering(863) 00:10:09.101 fused_ordering(864) 00:10:09.101 fused_ordering(865) 00:10:09.101 fused_ordering(866) 00:10:09.101 fused_ordering(867) 00:10:09.101 fused_ordering(868) 00:10:09.101 fused_ordering(869) 00:10:09.101 fused_ordering(870) 00:10:09.101 fused_ordering(871) 00:10:09.101 fused_ordering(872) 00:10:09.101 fused_ordering(873) 00:10:09.101 fused_ordering(874) 00:10:09.101 fused_ordering(875) 00:10:09.101 fused_ordering(876) 00:10:09.101 fused_ordering(877) 00:10:09.101 fused_ordering(878) 00:10:09.101 fused_ordering(879) 00:10:09.101 fused_ordering(880) 00:10:09.101 fused_ordering(881) 00:10:09.101 fused_ordering(882) 00:10:09.101 fused_ordering(883) 00:10:09.101 fused_ordering(884) 00:10:09.101 fused_ordering(885) 00:10:09.101 fused_ordering(886) 00:10:09.101 fused_ordering(887) 00:10:09.101 fused_ordering(888) 00:10:09.101 fused_ordering(889) 00:10:09.101 fused_ordering(890) 00:10:09.101 fused_ordering(891) 00:10:09.101 fused_ordering(892) 00:10:09.101 fused_ordering(893) 00:10:09.101 fused_ordering(894) 00:10:09.101 fused_ordering(895) 00:10:09.101 fused_ordering(896) 00:10:09.101 fused_ordering(897) 00:10:09.101 fused_ordering(898) 00:10:09.101 fused_ordering(899) 00:10:09.101 fused_ordering(900) 00:10:09.101 fused_ordering(901) 00:10:09.101 fused_ordering(902) 00:10:09.101 fused_ordering(903) 00:10:09.101 fused_ordering(904) 00:10:09.101 fused_ordering(905) 00:10:09.101 fused_ordering(906) 00:10:09.101 fused_ordering(907) 00:10:09.101 fused_ordering(908) 00:10:09.101 fused_ordering(909) 00:10:09.101 fused_ordering(910) 00:10:09.101 fused_ordering(911) 00:10:09.101 fused_ordering(912) 00:10:09.101 fused_ordering(913) 00:10:09.101 fused_ordering(914) 00:10:09.101 fused_ordering(915) 00:10:09.101 fused_ordering(916) 00:10:09.101 fused_ordering(917) 00:10:09.101 fused_ordering(918) 00:10:09.101 fused_ordering(919) 00:10:09.101 fused_ordering(920) 00:10:09.101 fused_ordering(921) 00:10:09.101 fused_ordering(922) 00:10:09.101 fused_ordering(923) 00:10:09.101 fused_ordering(924) 00:10:09.101 fused_ordering(925) 00:10:09.101 fused_ordering(926) 00:10:09.101 fused_ordering(927) 00:10:09.101 fused_ordering(928) 00:10:09.101 fused_ordering(929) 00:10:09.101 fused_ordering(930) 00:10:09.101 fused_ordering(931) 00:10:09.101 fused_ordering(932) 00:10:09.101 fused_ordering(933) 00:10:09.101 fused_ordering(934) 00:10:09.101 fused_ordering(935) 00:10:09.101 fused_ordering(936) 00:10:09.101 fused_ordering(937) 00:10:09.101 fused_ordering(938) 00:10:09.101 fused_ordering(939) 00:10:09.101 fused_ordering(940) 00:10:09.101 fused_ordering(941) 00:10:09.101 fused_ordering(942) 00:10:09.101 fused_ordering(943) 00:10:09.101 fused_ordering(944) 00:10:09.101 fused_ordering(945) 00:10:09.101 fused_ordering(946) 00:10:09.101 fused_ordering(947) 00:10:09.101 fused_ordering(948) 00:10:09.101 fused_ordering(949) 00:10:09.101 fused_ordering(950) 00:10:09.101 fused_ordering(951) 00:10:09.101 fused_ordering(952) 00:10:09.101 fused_ordering(953) 00:10:09.101 fused_ordering(954) 00:10:09.101 fused_ordering(955) 00:10:09.101 fused_ordering(956) 00:10:09.101 fused_ordering(957) 00:10:09.101 fused_ordering(958) 00:10:09.101 fused_ordering(959) 00:10:09.101 fused_ordering(960) 00:10:09.101 fused_ordering(961) 00:10:09.101 fused_ordering(962) 00:10:09.101 fused_ordering(963) 00:10:09.101 fused_ordering(964) 00:10:09.101 fused_ordering(965) 00:10:09.101 fused_ordering(966) 00:10:09.101 fused_ordering(967) 00:10:09.101 fused_ordering(968) 00:10:09.101 fused_ordering(969) 00:10:09.101 fused_ordering(970) 00:10:09.101 fused_ordering(971) 00:10:09.101 fused_ordering(972) 00:10:09.101 fused_ordering(973) 00:10:09.101 fused_ordering(974) 00:10:09.101 fused_ordering(975) 00:10:09.101 fused_ordering(976) 00:10:09.101 fused_ordering(977) 00:10:09.101 fused_ordering(978) 00:10:09.101 fused_ordering(979) 00:10:09.101 fused_ordering(980) 00:10:09.101 fused_ordering(981) 00:10:09.101 fused_ordering(982) 00:10:09.101 fused_ordering(983) 00:10:09.101 fused_ordering(984) 00:10:09.101 fused_ordering(985) 00:10:09.101 fused_ordering(986) 00:10:09.101 fused_ordering(987) 00:10:09.101 fused_ordering(988) 00:10:09.101 fused_ordering(989) 00:10:09.101 fused_ordering(990) 00:10:09.101 fused_ordering(991) 00:10:09.101 fused_ordering(992) 00:10:09.101 fused_ordering(993) 00:10:09.101 fused_ordering(994) 00:10:09.101 fused_ordering(995) 00:10:09.101 fused_ordering(996) 00:10:09.101 fused_ordering(997) 00:10:09.101 fused_ordering(998) 00:10:09.101 fused_ordering(999) 00:10:09.101 fused_ordering(1000) 00:10:09.101 fused_ordering(1001) 00:10:09.101 fused_ordering(1002) 00:10:09.101 fused_ordering(1003) 00:10:09.101 fused_ordering(1004) 00:10:09.101 fused_ordering(1005) 00:10:09.101 fused_ordering(1006) 00:10:09.101 fused_ordering(1007) 00:10:09.101 fused_ordering(1008) 00:10:09.101 fused_ordering(1009) 00:10:09.101 fused_ordering(1010) 00:10:09.101 fused_ordering(1011) 00:10:09.101 fused_ordering(1012) 00:10:09.101 fused_ordering(1013) 00:10:09.101 fused_ordering(1014) 00:10:09.101 fused_ordering(1015) 00:10:09.101 fused_ordering(1016) 00:10:09.101 fused_ordering(1017) 00:10:09.101 fused_ordering(1018) 00:10:09.101 fused_ordering(1019) 00:10:09.101 fused_ordering(1020) 00:10:09.101 fused_ordering(1021) 00:10:09.101 fused_ordering(1022) 00:10:09.101 fused_ordering(1023) 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- target/fused_ordering.sh@23 -- # trap - SIGINT SIGTERM EXIT 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- target/fused_ordering.sh@25 -- # nvmftestfini 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@488 -- # nvmfcleanup 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@117 -- # sync 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@120 -- # set +e 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@121 -- # for i in {1..20} 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:10:09.101 rmmod nvme_tcp 00:10:09.101 rmmod nvme_fabrics 00:10:09.101 rmmod nvme_keyring 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@124 -- # set -e 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@125 -- # return 0 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@489 -- # '[' -n 493074 ']' 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@490 -- # killprocess 493074 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@948 -- # '[' -z 493074 ']' 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@952 -- # kill -0 493074 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@953 -- # uname 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 493074 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@966 -- # echo 'killing process with pid 493074' 00:10:09.101 killing process with pid 493074 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@967 -- # kill 493074 00:10:09.101 17:56:02 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@972 -- # wait 493074 00:10:09.361 17:56:02 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:10:09.361 17:56:02 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:10:09.361 17:56:02 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:10:09.361 17:56:02 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:10:09.361 17:56:02 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@278 -- # remove_spdk_ns 00:10:09.361 17:56:02 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:09.361 17:56:02 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:10:09.361 17:56:02 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:11.267 17:56:04 nvmf_tcp.nvmf_fused_ordering -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:10:11.267 00:10:11.267 real 0m10.717s 00:10:11.267 user 0m5.504s 00:10:11.267 sys 0m5.552s 00:10:11.267 17:56:04 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@1124 -- # xtrace_disable 00:10:11.267 17:56:04 nvmf_tcp.nvmf_fused_ordering -- common/autotest_common.sh@10 -- # set +x 00:10:11.267 ************************************ 00:10:11.267 END TEST nvmf_fused_ordering 00:10:11.267 ************************************ 00:10:11.267 17:56:04 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:10:11.267 17:56:04 nvmf_tcp -- nvmf/nvmf.sh@35 -- # run_test nvmf_delete_subsystem /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh --transport=tcp 00:10:11.267 17:56:04 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:10:11.267 17:56:04 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:10:11.267 17:56:04 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:10:11.525 ************************************ 00:10:11.525 START TEST nvmf_delete_subsystem 00:10:11.525 ************************************ 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh --transport=tcp 00:10:11.525 * Looking for test storage... 00:10:11.525 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@7 -- # uname -s 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- paths/export.sh@5 -- # export PATH 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@47 -- # : 0 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:10:11.525 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:11.526 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:11.526 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:11.526 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:10:11.526 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:10:11.526 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@51 -- # have_pci_nics=0 00:10:11.526 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@12 -- # nvmftestinit 00:10:11.526 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:10:11.526 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:10:11.526 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@448 -- # prepare_net_devs 00:10:11.526 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@410 -- # local -g is_hw=no 00:10:11.526 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@412 -- # remove_spdk_ns 00:10:11.526 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:11.526 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:10:11.526 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:11.526 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:10:11.526 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:10:11.526 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@285 -- # xtrace_disable 00:10:11.526 17:56:05 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@291 -- # pci_devs=() 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@291 -- # local -a pci_devs 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@292 -- # pci_net_devs=() 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@293 -- # pci_drivers=() 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@293 -- # local -A pci_drivers 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@295 -- # net_devs=() 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@295 -- # local -ga net_devs 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@296 -- # e810=() 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@296 -- # local -ga e810 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@297 -- # x722=() 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@297 -- # local -ga x722 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@298 -- # mlx=() 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@298 -- # local -ga mlx 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:10:16.824 Found 0000:86:00.0 (0x8086 - 0x159b) 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:10:16.824 Found 0000:86:00.1 (0x8086 - 0x159b) 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:10:16.824 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@390 -- # [[ up == up ]] 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:10:16.825 Found net devices under 0000:86:00.0: cvl_0_0 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@390 -- # [[ up == up ]] 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:10:16.825 Found net devices under 0000:86:00.1: cvl_0_1 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:10:16.825 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@414 -- # is_hw=yes 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:10:17.084 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:10:17.084 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.150 ms 00:10:17.084 00:10:17.084 --- 10.0.0.2 ping statistics --- 00:10:17.084 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:17.084 rtt min/avg/max/mdev = 0.150/0.150/0.150/0.000 ms 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:10:17.084 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:10:17.084 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.078 ms 00:10:17.084 00:10:17.084 --- 10.0.0.1 ping statistics --- 00:10:17.084 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:17.084 rtt min/avg/max/mdev = 0.078/0.078/0.078/0.000 ms 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@422 -- # return 0 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:10:17.084 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:10:17.345 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@13 -- # nvmfappstart -m 0x3 00:10:17.345 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:10:17.345 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@722 -- # xtrace_disable 00:10:17.345 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:17.345 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@481 -- # nvmfpid=497075 00:10:17.345 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x3 00:10:17.345 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@482 -- # waitforlisten 497075 00:10:17.345 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@829 -- # '[' -z 497075 ']' 00:10:17.345 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:17.345 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@834 -- # local max_retries=100 00:10:17.345 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:17.345 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:17.345 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@838 -- # xtrace_disable 00:10:17.345 17:56:10 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:17.345 [2024-07-15 17:56:10.871642] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:10:17.345 [2024-07-15 17:56:10.871688] [ DPDK EAL parameters: nvmf -c 0x3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:17.345 EAL: No free 2048 kB hugepages reported on node 1 00:10:17.345 [2024-07-15 17:56:10.930082] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:10:17.345 [2024-07-15 17:56:11.001941] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:10:17.345 [2024-07-15 17:56:11.001983] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:10:17.345 [2024-07-15 17:56:11.001990] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:10:17.345 [2024-07-15 17:56:11.001996] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:10:17.345 [2024-07-15 17:56:11.002001] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:10:17.345 [2024-07-15 17:56:11.002045] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:10:17.345 [2024-07-15 17:56:11.002048] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:10:17.993 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:10:17.993 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@862 -- # return 0 00:10:17.993 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:10:17.993 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@728 -- # xtrace_disable 00:10:17.993 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:17.993 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:10:17.993 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@15 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:10:17.993 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:17.993 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:17.994 [2024-07-15 17:56:11.713898] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:10:17.994 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:17.994 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@16 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:10:17.994 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:17.994 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:18.253 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:18.253 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@17 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:10:18.253 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:18.253 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:18.253 [2024-07-15 17:56:11.734076] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:10:18.253 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:18.253 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@18 -- # rpc_cmd bdev_null_create NULL1 1000 512 00:10:18.253 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:18.253 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:18.253 NULL1 00:10:18.253 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:18.253 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@23 -- # rpc_cmd bdev_delay_create -b NULL1 -d Delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:10:18.253 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:18.253 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:18.253 Delay0 00:10:18.253 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:18.253 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:18.253 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:18.253 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:18.253 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:18.253 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@28 -- # perf_pid=497308 00:10:18.253 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@30 -- # sleep 2 00:10:18.253 17:56:11 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 5 -q 128 -w randrw -M 70 -o 512 -P 4 00:10:18.253 EAL: No free 2048 kB hugepages reported on node 1 00:10:18.253 [2024-07-15 17:56:11.814995] subsystem.c:1568:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:10:20.155 17:56:13 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@32 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:10:20.155 17:56:13 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:20.155 17:56:13 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 starting I/O failed: -6 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 starting I/O failed: -6 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 starting I/O failed: -6 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 starting I/O failed: -6 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 starting I/O failed: -6 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 starting I/O failed: -6 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 starting I/O failed: -6 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 starting I/O failed: -6 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 starting I/O failed: -6 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 starting I/O failed: -6 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 [2024-07-15 17:56:13.947102] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7f50ec000c00 is same with the state(5) to be set 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Write completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.414 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 starting I/O failed: -6 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 starting I/O failed: -6 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 starting I/O failed: -6 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 starting I/O failed: -6 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 starting I/O failed: -6 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 starting I/O failed: -6 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 starting I/O failed: -6 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 starting I/O failed: -6 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 starting I/O failed: -6 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 starting I/O failed: -6 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 starting I/O failed: -6 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Write completed with error (sct=0, sc=8) 00:10:20.415 starting I/O failed: -6 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 Read completed with error (sct=0, sc=8) 00:10:20.415 [2024-07-15 17:56:13.947846] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20835c0 is same with the state(5) to be set 00:10:21.353 [2024-07-15 17:56:14.910652] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2084ac0 is same with the state(5) to be set 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 [2024-07-15 17:56:14.949372] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2083000 is same with the state(5) to be set 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 [2024-07-15 17:56:14.949548] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20837a0 is same with the state(5) to be set 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Write completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.353 Read completed with error (sct=0, sc=8) 00:10:21.354 Read completed with error (sct=0, sc=8) 00:10:21.354 [2024-07-15 17:56:14.949711] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x20833e0 is same with the state(5) to be set 00:10:21.354 Read completed with error (sct=0, sc=8) 00:10:21.354 Read completed with error (sct=0, sc=8) 00:10:21.354 Read completed with error (sct=0, sc=8) 00:10:21.354 Write completed with error (sct=0, sc=8) 00:10:21.354 Write completed with error (sct=0, sc=8) 00:10:21.354 Read completed with error (sct=0, sc=8) 00:10:21.354 Read completed with error (sct=0, sc=8) 00:10:21.354 Read completed with error (sct=0, sc=8) 00:10:21.354 Read completed with error (sct=0, sc=8) 00:10:21.354 Read completed with error (sct=0, sc=8) 00:10:21.354 Read completed with error (sct=0, sc=8) 00:10:21.354 Write completed with error (sct=0, sc=8) 00:10:21.354 Read completed with error (sct=0, sc=8) 00:10:21.354 Read completed with error (sct=0, sc=8) 00:10:21.354 Read completed with error (sct=0, sc=8) 00:10:21.354 Write completed with error (sct=0, sc=8) 00:10:21.354 Read completed with error (sct=0, sc=8) 00:10:21.354 Read completed with error (sct=0, sc=8) 00:10:21.354 Read completed with error (sct=0, sc=8) 00:10:21.354 Read completed with error (sct=0, sc=8) 00:10:21.354 Write completed with error (sct=0, sc=8) 00:10:21.354 [2024-07-15 17:56:14.949818] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x7f50ec00d2f0 is same with the state(5) to be set 00:10:21.354 Initializing NVMe Controllers 00:10:21.354 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:10:21.354 Controller IO queue size 128, less than required. 00:10:21.354 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:10:21.354 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:10:21.354 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:10:21.354 Initialization complete. Launching workers. 00:10:21.354 ======================================================== 00:10:21.354 Latency(us) 00:10:21.354 Device Information : IOPS MiB/s Average min max 00:10:21.354 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 179.25 0.09 1035104.10 1525.01 2001578.38 00:10:21.354 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 154.92 0.08 893278.84 235.62 2001800.17 00:10:21.354 ======================================================== 00:10:21.354 Total : 334.17 0.16 969354.50 235.62 2001800.17 00:10:21.354 00:10:21.354 [2024-07-15 17:56:14.950413] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x2084ac0 (9): Bad file descriptor 00:10:21.354 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf: errors occurred 00:10:21.354 17:56:14 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:21.354 17:56:14 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@34 -- # delay=0 00:10:21.354 17:56:14 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@35 -- # kill -0 497308 00:10:21.354 17:56:14 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@36 -- # sleep 0.5 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@38 -- # (( delay++ > 30 )) 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@35 -- # kill -0 497308 00:10:21.922 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh: line 35: kill: (497308) - No such process 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@45 -- # NOT wait 497308 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@648 -- # local es=0 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@650 -- # valid_exec_arg wait 497308 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@636 -- # local arg=wait 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@640 -- # type -t wait 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@651 -- # wait 497308 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@651 -- # es=1 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@48 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@49 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:21.922 [2024-07-15 17:56:15.480772] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@50 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Delay0 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@54 -- # perf_pid=497988 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@56 -- # delay=0 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -c 0xC -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -t 3 -q 128 -w randrw -M 70 -o 512 -P 4 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 497988 00:10:21.922 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:10:21.922 EAL: No free 2048 kB hugepages reported on node 1 00:10:21.922 [2024-07-15 17:56:15.540211] subsystem.c:1568:spdk_nvmf_subsystem_listener_allowed: *WARNING*: Allowing connection to discovery subsystem on TCP/10.0.0.2/4420, even though this listener was not added to the discovery subsystem. This behavior is deprecated and will be removed in a future release. 00:10:22.491 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:10:22.491 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 497988 00:10:22.491 17:56:15 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:10:23.060 17:56:16 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:10:23.060 17:56:16 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 497988 00:10:23.060 17:56:16 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:10:23.318 17:56:17 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:10:23.318 17:56:17 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 497988 00:10:23.318 17:56:17 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:10:23.887 17:56:17 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:10:23.887 17:56:17 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 497988 00:10:23.887 17:56:17 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:10:24.454 17:56:18 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:10:24.454 17:56:18 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 497988 00:10:24.454 17:56:18 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:10:25.022 17:56:18 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:10:25.022 17:56:18 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 497988 00:10:25.022 17:56:18 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@58 -- # sleep 0.5 00:10:25.281 Initializing NVMe Controllers 00:10:25.281 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:10:25.281 Controller IO queue size 128, less than required. 00:10:25.281 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:10:25.281 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 2 00:10:25.281 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 3 00:10:25.281 Initialization complete. Launching workers. 00:10:25.281 ======================================================== 00:10:25.281 Latency(us) 00:10:25.281 Device Information : IOPS MiB/s Average min max 00:10:25.281 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 2: 128.00 0.06 1003659.94 1000191.87 1040801.73 00:10:25.281 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 3: 128.00 0.06 1005053.77 1000294.55 1041560.07 00:10:25.281 ======================================================== 00:10:25.281 Total : 256.00 0.12 1004356.85 1000191.87 1041560.07 00:10:25.281 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@60 -- # (( delay++ > 20 )) 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@57 -- # kill -0 497988 00:10:25.540 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/delete_subsystem.sh: line 57: kill: (497988) - No such process 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@67 -- # wait 497988 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@69 -- # trap - SIGINT SIGTERM EXIT 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- target/delete_subsystem.sh@71 -- # nvmftestfini 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@488 -- # nvmfcleanup 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@117 -- # sync 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@120 -- # set +e 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@121 -- # for i in {1..20} 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:10:25.540 rmmod nvme_tcp 00:10:25.540 rmmod nvme_fabrics 00:10:25.540 rmmod nvme_keyring 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@124 -- # set -e 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@125 -- # return 0 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@489 -- # '[' -n 497075 ']' 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@490 -- # killprocess 497075 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@948 -- # '[' -z 497075 ']' 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@952 -- # kill -0 497075 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@953 -- # uname 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 497075 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@966 -- # echo 'killing process with pid 497075' 00:10:25.540 killing process with pid 497075 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@967 -- # kill 497075 00:10:25.540 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@972 -- # wait 497075 00:10:25.799 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:10:25.799 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:10:25.799 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:10:25.799 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:10:25.799 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@278 -- # remove_spdk_ns 00:10:25.799 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:25.799 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:10:25.799 17:56:19 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:27.705 17:56:21 nvmf_tcp.nvmf_delete_subsystem -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:10:27.705 00:10:27.705 real 0m16.384s 00:10:27.705 user 0m30.479s 00:10:27.705 sys 0m5.163s 00:10:27.705 17:56:21 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@1124 -- # xtrace_disable 00:10:27.705 17:56:21 nvmf_tcp.nvmf_delete_subsystem -- common/autotest_common.sh@10 -- # set +x 00:10:27.705 ************************************ 00:10:27.705 END TEST nvmf_delete_subsystem 00:10:27.705 ************************************ 00:10:27.705 17:56:21 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:10:27.705 17:56:21 nvmf_tcp -- nvmf/nvmf.sh@36 -- # run_test nvmf_ns_masking test/nvmf/target/ns_masking.sh --transport=tcp 00:10:27.705 17:56:21 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:10:27.705 17:56:21 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:10:27.705 17:56:21 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:10:27.967 ************************************ 00:10:27.967 START TEST nvmf_ns_masking 00:10:27.967 ************************************ 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1123 -- # test/nvmf/target/ns_masking.sh --transport=tcp 00:10:27.967 * Looking for test storage... 00:10:27.967 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@8 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@7 -- # uname -s 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- paths/export.sh@5 -- # export PATH 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@47 -- # : 0 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@51 -- # have_pci_nics=0 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@10 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@11 -- # hostsock=/var/tmp/host.sock 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@12 -- # loops=5 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@13 -- # uuidgen 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@13 -- # ns1uuid=e1317105-fa08-4905-8665-8d5b7621be69 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@14 -- # uuidgen 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@14 -- # ns2uuid=88c028b3-007d-4829-ad8f-b057881347c1 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@16 -- # SUBSYSNQN=nqn.2016-06.io.spdk:cnode1 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@17 -- # HOSTNQN1=nqn.2016-06.io.spdk:host1 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@18 -- # HOSTNQN2=nqn.2016-06.io.spdk:host2 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@19 -- # uuidgen 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@19 -- # HOSTID=764aeae3-6ca4-4c8a-b6e8-7c8914800492 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@50 -- # nvmftestinit 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@448 -- # prepare_net_devs 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@410 -- # local -g is_hw=no 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@412 -- # remove_spdk_ns 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@285 -- # xtrace_disable 00:10:27.967 17:56:21 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@291 -- # pci_devs=() 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@291 -- # local -a pci_devs 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@292 -- # pci_net_devs=() 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@293 -- # pci_drivers=() 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@293 -- # local -A pci_drivers 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@295 -- # net_devs=() 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@295 -- # local -ga net_devs 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@296 -- # e810=() 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@296 -- # local -ga e810 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@297 -- # x722=() 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@297 -- # local -ga x722 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@298 -- # mlx=() 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@298 -- # local -ga mlx 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:10:33.264 Found 0000:86:00.0 (0x8086 - 0x159b) 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:10:33.264 Found 0000:86:00.1 (0x8086 - 0x159b) 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@390 -- # [[ up == up ]] 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:10:33.264 Found net devices under 0000:86:00.0: cvl_0_0 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:33.264 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@390 -- # [[ up == up ]] 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:10:33.265 Found net devices under 0000:86:00.1: cvl_0_1 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@414 -- # is_hw=yes 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:10:33.265 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:10:33.265 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.277 ms 00:10:33.265 00:10:33.265 --- 10.0.0.2 ping statistics --- 00:10:33.265 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:33.265 rtt min/avg/max/mdev = 0.277/0.277/0.277/0.000 ms 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:10:33.265 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:10:33.265 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.098 ms 00:10:33.265 00:10:33.265 --- 10.0.0.1 ping statistics --- 00:10:33.265 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:33.265 rtt min/avg/max/mdev = 0.098/0.098/0.098/0.000 ms 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@422 -- # return 0 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@51 -- # nvmfappstart 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@722 -- # xtrace_disable 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@481 -- # nvmfpid=501938 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@482 -- # waitforlisten 501938 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@829 -- # '[' -z 501938 ']' 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@834 -- # local max_retries=100 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:33.265 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@838 -- # xtrace_disable 00:10:33.265 17:56:26 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:10:33.265 [2024-07-15 17:56:26.361787] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:10:33.265 [2024-07-15 17:56:26.361830] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:33.265 EAL: No free 2048 kB hugepages reported on node 1 00:10:33.265 [2024-07-15 17:56:26.418992] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:33.265 [2024-07-15 17:56:26.498112] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:10:33.265 [2024-07-15 17:56:26.498145] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:10:33.265 [2024-07-15 17:56:26.498152] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:10:33.265 [2024-07-15 17:56:26.498158] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:10:33.265 [2024-07-15 17:56:26.498163] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:10:33.265 [2024-07-15 17:56:26.498184] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:10:33.522 17:56:27 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:10:33.522 17:56:27 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@862 -- # return 0 00:10:33.522 17:56:27 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:10:33.522 17:56:27 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@728 -- # xtrace_disable 00:10:33.522 17:56:27 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:10:33.522 17:56:27 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:10:33.522 17:56:27 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:10:33.779 [2024-07-15 17:56:27.345761] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:10:33.779 17:56:27 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@55 -- # MALLOC_BDEV_SIZE=64 00:10:33.779 17:56:27 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@56 -- # MALLOC_BLOCK_SIZE=512 00:10:33.779 17:56:27 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:10:34.037 Malloc1 00:10:34.037 17:56:27 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc2 00:10:34.037 Malloc2 00:10:34.295 17:56:27 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@62 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:10:34.295 17:56:27 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 1 00:10:34.553 17:56:28 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:10:34.553 [2024-07-15 17:56:28.261910] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:10:34.810 17:56:28 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@67 -- # connect 00:10:34.810 17:56:28 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@22 -- # nvme connect -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -I 764aeae3-6ca4-4c8a-b6e8-7c8914800492 -a 10.0.0.2 -s 4420 -i 4 00:10:34.810 17:56:28 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@24 -- # waitforserial SPDKISFASTANDAWESOME 00:10:34.810 17:56:28 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1198 -- # local i=0 00:10:34.810 17:56:28 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:10:34.810 17:56:28 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:10:34.810 17:56:28 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1205 -- # sleep 2 00:10:36.709 17:56:30 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:10:36.709 17:56:30 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:10:36.709 17:56:30 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:10:36.709 17:56:30 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:10:36.709 17:56:30 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:10:36.709 17:56:30 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1208 -- # return 0 00:10:36.992 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@26 -- # jq -r '.[].Subsystems[] | select(.NQN=="nqn.2016-06.io.spdk:cnode1") | .Paths[0].Name' 00:10:36.992 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@26 -- # nvme list-subsys -o json 00:10:36.992 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@26 -- # ctrl_id=nvme0 00:10:36.992 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@27 -- # [[ -z nvme0 ]] 00:10:36.992 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@68 -- # ns_is_visible 0x1 00:10:36.992 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:10:36.992 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:10:36.992 [ 0]:0x1 00:10:36.992 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:10:36.992 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:10:36.992 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=22a0b9bba4ba461982989d35a79d3c77 00:10:36.992 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 22a0b9bba4ba461982989d35a79d3c77 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:10:36.992 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc2 -n 2 00:10:36.992 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@72 -- # ns_is_visible 0x1 00:10:36.992 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:10:36.992 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:10:36.992 [ 0]:0x1 00:10:36.992 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:10:36.992 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:10:37.250 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=22a0b9bba4ba461982989d35a79d3c77 00:10:37.251 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 22a0b9bba4ba461982989d35a79d3c77 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:10:37.251 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@73 -- # ns_is_visible 0x2 00:10:37.251 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:10:37.251 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:10:37.251 [ 1]:0x2 00:10:37.251 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:10:37.251 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:10:37.251 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=43c1a3c6fc1f4cdc8b956e6e5c1da809 00:10:37.251 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 43c1a3c6fc1f4cdc8b956e6e5c1da809 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:10:37.251 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@75 -- # disconnect 00:10:37.251 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:10:37.251 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:10:37.251 17:56:30 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:37.508 17:56:31 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 1 --no-auto-visible 00:10:37.508 17:56:31 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@83 -- # connect 1 00:10:37.508 17:56:31 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@22 -- # nvme connect -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -I 764aeae3-6ca4-4c8a-b6e8-7c8914800492 -a 10.0.0.2 -s 4420 -i 4 00:10:37.765 17:56:31 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@24 -- # waitforserial SPDKISFASTANDAWESOME 1 00:10:37.765 17:56:31 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1198 -- # local i=0 00:10:37.765 17:56:31 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:10:37.765 17:56:31 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1200 -- # [[ -n 1 ]] 00:10:37.765 17:56:31 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1201 -- # nvme_device_counter=1 00:10:37.765 17:56:31 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1205 -- # sleep 2 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1208 -- # return 0 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@26 -- # nvme list-subsys -o json 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@26 -- # jq -r '.[].Subsystems[] | select(.NQN=="nqn.2016-06.io.spdk:cnode1") | .Paths[0].Name' 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@26 -- # ctrl_id=nvme0 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@27 -- # [[ -z nvme0 ]] 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@84 -- # NOT ns_is_visible 0x1 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@648 -- # local es=0 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@650 -- # valid_exec_arg ns_is_visible 0x1 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@636 -- # local arg=ns_is_visible 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@640 -- # type -t ns_is_visible 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@651 -- # ns_is_visible 0x1 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=00000000000000000000000000000000 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@651 -- # es=1 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@85 -- # ns_is_visible 0x2 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:10:40.300 [ 0]:0x2 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=43c1a3c6fc1f4cdc8b956e6e5c1da809 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 43c1a3c6fc1f4cdc8b956e6e5c1da809 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@89 -- # ns_is_visible 0x1 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:10:40.300 [ 0]:0x1 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=22a0b9bba4ba461982989d35a79d3c77 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 22a0b9bba4ba461982989d35a79d3c77 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@90 -- # ns_is_visible 0x2 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:10:40.300 [ 1]:0x2 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=43c1a3c6fc1f4cdc8b956e6e5c1da809 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 43c1a3c6fc1f4cdc8b956e6e5c1da809 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:10:40.300 17:56:33 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@94 -- # NOT ns_is_visible 0x1 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@648 -- # local es=0 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@650 -- # valid_exec_arg ns_is_visible 0x1 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@636 -- # local arg=ns_is_visible 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@640 -- # type -t ns_is_visible 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@651 -- # ns_is_visible 0x1 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=00000000000000000000000000000000 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@651 -- # es=1 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@95 -- # ns_is_visible 0x2 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:10:40.559 [ 0]:0x2 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=43c1a3c6fc1f4cdc8b956e6e5c1da809 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 43c1a3c6fc1f4cdc8b956e6e5c1da809 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@97 -- # disconnect 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:10:40.559 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:10:40.559 17:56:34 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:10:40.818 17:56:34 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@101 -- # connect 2 00:10:40.818 17:56:34 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@22 -- # nvme connect -t tcp -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -I 764aeae3-6ca4-4c8a-b6e8-7c8914800492 -a 10.0.0.2 -s 4420 -i 4 00:10:40.818 17:56:34 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@24 -- # waitforserial SPDKISFASTANDAWESOME 2 00:10:40.818 17:56:34 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1198 -- # local i=0 00:10:40.818 17:56:34 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:10:40.818 17:56:34 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1200 -- # [[ -n 2 ]] 00:10:40.818 17:56:34 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1201 -- # nvme_device_counter=2 00:10:40.818 17:56:34 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1205 -- # sleep 2 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1207 -- # nvme_devices=2 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1208 -- # return 0 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@26 -- # nvme list-subsys -o json 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@26 -- # jq -r '.[].Subsystems[] | select(.NQN=="nqn.2016-06.io.spdk:cnode1") | .Paths[0].Name' 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@26 -- # ctrl_id=nvme0 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@27 -- # [[ -z nvme0 ]] 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@102 -- # ns_is_visible 0x1 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:10:43.356 [ 0]:0x1 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=22a0b9bba4ba461982989d35a79d3c77 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 22a0b9bba4ba461982989d35a79d3c77 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@103 -- # ns_is_visible 0x2 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:10:43.356 [ 1]:0x2 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=43c1a3c6fc1f4cdc8b956e6e5c1da809 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 43c1a3c6fc1f4cdc8b956e6e5c1da809 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@106 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@107 -- # NOT ns_is_visible 0x1 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@648 -- # local es=0 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@650 -- # valid_exec_arg ns_is_visible 0x1 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@636 -- # local arg=ns_is_visible 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@640 -- # type -t ns_is_visible 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@651 -- # ns_is_visible 0x1 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:10:43.356 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=00000000000000000000000000000000 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@651 -- # es=1 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@108 -- # ns_is_visible 0x2 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:10:43.357 [ 0]:0x2 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=43c1a3c6fc1f4cdc8b956e6e5c1da809 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 43c1a3c6fc1f4cdc8b956e6e5c1da809 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@111 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host1 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@648 -- # local es=0 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host1 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:10:43.357 17:56:36 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_remove_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host1 00:10:43.617 [2024-07-15 17:56:37.114766] nvmf_rpc.c:1791:nvmf_rpc_ns_visible_paused: *ERROR*: Unable to add/remove nqn.2016-06.io.spdk:host1 to namespace ID 2 00:10:43.617 request: 00:10:43.617 { 00:10:43.617 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:10:43.617 "nsid": 2, 00:10:43.617 "host": "nqn.2016-06.io.spdk:host1", 00:10:43.617 "method": "nvmf_ns_remove_host", 00:10:43.617 "req_id": 1 00:10:43.617 } 00:10:43.617 Got JSON-RPC error response 00:10:43.617 response: 00:10:43.617 { 00:10:43.617 "code": -32602, 00:10:43.617 "message": "Invalid parameters" 00:10:43.617 } 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@651 -- # es=1 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@112 -- # NOT ns_is_visible 0x1 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@648 -- # local es=0 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@650 -- # valid_exec_arg ns_is_visible 0x1 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@636 -- # local arg=ns_is_visible 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@640 -- # type -t ns_is_visible 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@651 -- # ns_is_visible 0x1 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x1 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x1 -o json 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=00000000000000000000000000000000 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 00000000000000000000000000000000 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@651 -- # es=1 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@113 -- # ns_is_visible 0x2 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # nvme list-ns /dev/nvme0 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@43 -- # grep 0x2 00:10:43.617 [ 0]:0x2 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nvme id-ns /dev/nvme0 -n 0x2 -o json 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # jq -r .nguid 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@44 -- # nguid=43c1a3c6fc1f4cdc8b956e6e5c1da809 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@45 -- # [[ 43c1a3c6fc1f4cdc8b956e6e5c1da809 != \0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0\0 ]] 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@114 -- # disconnect 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@38 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:10:43.617 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@118 -- # hostpid=503898 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@117 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -r /var/tmp/host.sock -m 2 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@119 -- # trap 'killprocess $hostpid; nvmftestfini' SIGINT SIGTERM EXIT 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@121 -- # waitforlisten 503898 /var/tmp/host.sock 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@829 -- # '[' -z 503898 ']' 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/host.sock 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@834 -- # local max_retries=100 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/host.sock...' 00:10:43.617 Waiting for process to start up and listen on UNIX domain socket /var/tmp/host.sock... 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@838 -- # xtrace_disable 00:10:43.617 17:56:37 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:10:43.876 [2024-07-15 17:56:37.345590] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:10:43.876 [2024-07-15 17:56:37.345637] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid503898 ] 00:10:43.876 EAL: No free 2048 kB hugepages reported on node 1 00:10:43.876 [2024-07-15 17:56:37.399425] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:10:43.876 [2024-07-15 17:56:37.471613] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:10:44.446 17:56:38 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:10:44.446 17:56:38 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@862 -- # return 0 00:10:44.446 17:56:38 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@122 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:10:44.704 17:56:38 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 2 00:10:44.963 17:56:38 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@124 -- # uuid2nguid e1317105-fa08-4905-8665-8d5b7621be69 00:10:44.963 17:56:38 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@759 -- # tr -d - 00:10:44.963 17:56:38 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@124 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 1 -g E1317105FA08490586658D5B7621BE69 -i 00:10:44.963 17:56:38 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@125 -- # uuid2nguid 88c028b3-007d-4829-ad8f-b057881347c1 00:10:44.963 17:56:38 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@759 -- # tr -d - 00:10:44.963 17:56:38 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@125 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc2 -n 2 -g 88C028B3007D4829AD8FB057881347C1 -i 00:10:45.222 17:56:38 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@126 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 1 nqn.2016-06.io.spdk:host1 00:10:45.482 17:56:39 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@127 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_ns_add_host nqn.2016-06.io.spdk:cnode1 2 nqn.2016-06.io.spdk:host2 00:10:45.482 17:56:39 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@129 -- # hostrpc bdev_nvme_attach_controller -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -b nvme0 00:10:45.482 17:56:39 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 -b nvme0 00:10:45.741 nvme0n1 00:10:45.741 17:56:39 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@131 -- # hostrpc bdev_nvme_attach_controller -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host2 -b nvme1 00:10:45.741 17:56:39 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host2 -b nvme1 00:10:46.311 nvme1n2 00:10:46.311 17:56:39 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@134 -- # hostrpc bdev_get_bdevs 00:10:46.311 17:56:39 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@134 -- # xargs 00:10:46.311 17:56:39 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@134 -- # jq -r '.[].name' 00:10:46.311 17:56:39 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_get_bdevs 00:10:46.311 17:56:39 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@134 -- # sort 00:10:46.311 17:56:39 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@134 -- # [[ nvme0n1 nvme1n2 == \n\v\m\e\0\n\1\ \n\v\m\e\1\n\2 ]] 00:10:46.311 17:56:40 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@135 -- # hostrpc bdev_get_bdevs -b nvme0n1 00:10:46.311 17:56:40 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@135 -- # jq -r '.[].uuid' 00:10:46.311 17:56:40 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_get_bdevs -b nvme0n1 00:10:46.570 17:56:40 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@135 -- # [[ e1317105-fa08-4905-8665-8d5b7621be69 == \e\1\3\1\7\1\0\5\-\f\a\0\8\-\4\9\0\5\-\8\6\6\5\-\8\d\5\b\7\6\2\1\b\e\6\9 ]] 00:10:46.570 17:56:40 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@136 -- # hostrpc bdev_get_bdevs -b nvme1n2 00:10:46.570 17:56:40 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@136 -- # jq -r '.[].uuid' 00:10:46.570 17:56:40 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_get_bdevs -b nvme1n2 00:10:46.829 17:56:40 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@136 -- # [[ 88c028b3-007d-4829-ad8f-b057881347c1 == \8\8\c\0\2\8\b\3\-\0\0\7\d\-\4\8\2\9\-\a\d\8\f\-\b\0\5\7\8\8\1\3\4\7\c\1 ]] 00:10:46.829 17:56:40 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@138 -- # killprocess 503898 00:10:46.829 17:56:40 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@948 -- # '[' -z 503898 ']' 00:10:46.829 17:56:40 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@952 -- # kill -0 503898 00:10:46.829 17:56:40 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@953 -- # uname 00:10:46.829 17:56:40 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:10:46.829 17:56:40 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 503898 00:10:46.829 17:56:40 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:10:46.829 17:56:40 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:10:46.829 17:56:40 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@966 -- # echo 'killing process with pid 503898' 00:10:46.829 killing process with pid 503898 00:10:46.829 17:56:40 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@967 -- # kill 503898 00:10:46.829 17:56:40 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@972 -- # wait 503898 00:10:47.089 17:56:40 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@139 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:10:47.348 17:56:40 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@141 -- # trap - SIGINT SIGTERM EXIT 00:10:47.348 17:56:40 nvmf_tcp.nvmf_ns_masking -- target/ns_masking.sh@142 -- # nvmftestfini 00:10:47.348 17:56:40 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@488 -- # nvmfcleanup 00:10:47.348 17:56:40 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@117 -- # sync 00:10:47.348 17:56:40 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:10:47.348 17:56:40 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@120 -- # set +e 00:10:47.348 17:56:40 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@121 -- # for i in {1..20} 00:10:47.348 17:56:40 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:10:47.348 rmmod nvme_tcp 00:10:47.348 rmmod nvme_fabrics 00:10:47.348 rmmod nvme_keyring 00:10:47.348 17:56:40 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:10:47.348 17:56:40 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@124 -- # set -e 00:10:47.348 17:56:40 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@125 -- # return 0 00:10:47.348 17:56:40 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@489 -- # '[' -n 501938 ']' 00:10:47.348 17:56:40 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@490 -- # killprocess 501938 00:10:47.348 17:56:40 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@948 -- # '[' -z 501938 ']' 00:10:47.348 17:56:40 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@952 -- # kill -0 501938 00:10:47.348 17:56:40 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@953 -- # uname 00:10:47.348 17:56:40 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:10:47.348 17:56:40 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 501938 00:10:47.348 17:56:41 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:10:47.348 17:56:41 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:10:47.348 17:56:41 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@966 -- # echo 'killing process with pid 501938' 00:10:47.348 killing process with pid 501938 00:10:47.348 17:56:41 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@967 -- # kill 501938 00:10:47.348 17:56:41 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@972 -- # wait 501938 00:10:47.607 17:56:41 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:10:47.607 17:56:41 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:10:47.607 17:56:41 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:10:47.607 17:56:41 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:10:47.607 17:56:41 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@278 -- # remove_spdk_ns 00:10:47.607 17:56:41 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:47.607 17:56:41 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:10:47.607 17:56:41 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:50.146 17:56:43 nvmf_tcp.nvmf_ns_masking -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:10:50.146 00:10:50.146 real 0m21.840s 00:10:50.146 user 0m23.948s 00:10:50.146 sys 0m5.631s 00:10:50.146 17:56:43 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@1124 -- # xtrace_disable 00:10:50.146 17:56:43 nvmf_tcp.nvmf_ns_masking -- common/autotest_common.sh@10 -- # set +x 00:10:50.146 ************************************ 00:10:50.146 END TEST nvmf_ns_masking 00:10:50.146 ************************************ 00:10:50.146 17:56:43 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:10:50.146 17:56:43 nvmf_tcp -- nvmf/nvmf.sh@37 -- # [[ 1 -eq 1 ]] 00:10:50.146 17:56:43 nvmf_tcp -- nvmf/nvmf.sh@38 -- # run_test nvmf_nvme_cli /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvme_cli.sh --transport=tcp 00:10:50.146 17:56:43 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:10:50.146 17:56:43 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:10:50.146 17:56:43 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:10:50.146 ************************************ 00:10:50.146 START TEST nvmf_nvme_cli 00:10:50.146 ************************************ 00:10:50.146 17:56:43 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvme_cli.sh --transport=tcp 00:10:50.146 * Looking for test storage... 00:10:50.146 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:10:50.146 17:56:43 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:10:50.146 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@7 -- # uname -s 00:10:50.146 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- paths/export.sh@5 -- # export PATH 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@47 -- # : 0 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@51 -- # have_pci_nics=0 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@11 -- # MALLOC_BDEV_SIZE=64 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@14 -- # devs=() 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@16 -- # nvmftestinit 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@448 -- # prepare_net_devs 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@410 -- # local -g is_hw=no 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@412 -- # remove_spdk_ns 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@285 -- # xtrace_disable 00:10:50.147 17:56:43 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@291 -- # pci_devs=() 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@291 -- # local -a pci_devs 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@292 -- # pci_net_devs=() 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@293 -- # pci_drivers=() 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@293 -- # local -A pci_drivers 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@295 -- # net_devs=() 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@295 -- # local -ga net_devs 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@296 -- # e810=() 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@296 -- # local -ga e810 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@297 -- # x722=() 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@297 -- # local -ga x722 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@298 -- # mlx=() 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@298 -- # local -ga mlx 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:10:55.447 Found 0000:86:00.0 (0x8086 - 0x159b) 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:10:55.447 Found 0000:86:00.1 (0x8086 - 0x159b) 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@390 -- # [[ up == up ]] 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:10:55.447 Found net devices under 0000:86:00.0: cvl_0_0 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@390 -- # [[ up == up ]] 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:10:55.447 Found net devices under 0000:86:00.1: cvl_0_1 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:10:55.447 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@414 -- # is_hw=yes 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:10:55.448 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:10:55.448 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.189 ms 00:10:55.448 00:10:55.448 --- 10.0.0.2 ping statistics --- 00:10:55.448 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:55.448 rtt min/avg/max/mdev = 0.189/0.189/0.189/0.000 ms 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:10:55.448 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:10:55.448 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.115 ms 00:10:55.448 00:10:55.448 --- 10.0.0.1 ping statistics --- 00:10:55.448 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:10:55.448 rtt min/avg/max/mdev = 0.115/0.115/0.115/0.000 ms 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@422 -- # return 0 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@17 -- # nvmfappstart -m 0xF 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@722 -- # xtrace_disable 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@481 -- # nvmfpid=507992 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@482 -- # waitforlisten 507992 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@829 -- # '[' -z 507992 ']' 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@834 -- # local max_retries=100 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:10:55.448 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@838 -- # xtrace_disable 00:10:55.448 17:56:48 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:10:55.448 [2024-07-15 17:56:48.713945] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:10:55.448 [2024-07-15 17:56:48.713991] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:10:55.448 EAL: No free 2048 kB hugepages reported on node 1 00:10:55.448 [2024-07-15 17:56:48.771518] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:10:55.448 [2024-07-15 17:56:48.854017] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:10:55.448 [2024-07-15 17:56:48.854053] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:10:55.448 [2024-07-15 17:56:48.854061] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:10:55.448 [2024-07-15 17:56:48.854066] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:10:55.448 [2024-07-15 17:56:48.854072] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:10:55.448 [2024-07-15 17:56:48.854114] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:10:55.448 [2024-07-15 17:56:48.854209] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:10:55.448 [2024-07-15 17:56:48.854294] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:10:55.448 [2024-07-15 17:56:48.854296] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@862 -- # return 0 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@728 -- # xtrace_disable 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@19 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:10:56.019 [2024-07-15 17:56:49.574244] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@21 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:10:56.019 Malloc0 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:10:56.019 Malloc1 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME -d SPDK_Controller1 -i 291 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@25 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@26 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:10:56.019 [2024-07-15 17:56:49.655520] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@28 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:56.019 17:56:49 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@30 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -a 10.0.0.2 -s 4420 00:10:56.279 00:10:56.279 Discovery Log Number of Records 2, Generation counter 2 00:10:56.279 =====Discovery Log Entry 0====== 00:10:56.279 trtype: tcp 00:10:56.279 adrfam: ipv4 00:10:56.279 subtype: current discovery subsystem 00:10:56.279 treq: not required 00:10:56.279 portid: 0 00:10:56.279 trsvcid: 4420 00:10:56.279 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:10:56.279 traddr: 10.0.0.2 00:10:56.279 eflags: explicit discovery connections, duplicate discovery information 00:10:56.279 sectype: none 00:10:56.279 =====Discovery Log Entry 1====== 00:10:56.279 trtype: tcp 00:10:56.279 adrfam: ipv4 00:10:56.279 subtype: nvme subsystem 00:10:56.279 treq: not required 00:10:56.279 portid: 0 00:10:56.279 trsvcid: 4420 00:10:56.279 subnqn: nqn.2016-06.io.spdk:cnode1 00:10:56.279 traddr: 10.0.0.2 00:10:56.279 eflags: none 00:10:56.279 sectype: none 00:10:56.279 17:56:49 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@31 -- # devs=($(get_nvme_devs)) 00:10:56.279 17:56:49 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@31 -- # get_nvme_devs 00:10:56.279 17:56:49 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@522 -- # local dev _ 00:10:56.279 17:56:49 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@524 -- # read -r dev _ 00:10:56.279 17:56:49 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@521 -- # nvme list 00:10:56.279 17:56:49 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@525 -- # [[ Node == /dev/nvme* ]] 00:10:56.279 17:56:49 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@524 -- # read -r dev _ 00:10:56.279 17:56:49 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@525 -- # [[ --------------------- == /dev/nvme* ]] 00:10:56.279 17:56:49 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@524 -- # read -r dev _ 00:10:56.279 17:56:49 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@31 -- # nvme_num_before_connection=0 00:10:56.279 17:56:49 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@32 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:10:57.690 17:56:50 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@34 -- # waitforserial SPDKISFASTANDAWESOME 2 00:10:57.690 17:56:50 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@1198 -- # local i=0 00:10:57.690 17:56:50 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:10:57.690 17:56:50 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@1200 -- # [[ -n 2 ]] 00:10:57.690 17:56:50 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@1201 -- # nvme_device_counter=2 00:10:57.690 17:56:50 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@1205 -- # sleep 2 00:10:59.591 17:56:52 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:10:59.591 17:56:52 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:10:59.591 17:56:52 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@1207 -- # nvme_devices=2 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@1208 -- # return 0 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@35 -- # get_nvme_devs 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@522 -- # local dev _ 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@524 -- # read -r dev _ 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@521 -- # nvme list 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@525 -- # [[ Node == /dev/nvme* ]] 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@524 -- # read -r dev _ 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@525 -- # [[ --------------------- == /dev/nvme* ]] 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@524 -- # read -r dev _ 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@525 -- # [[ /dev/nvme0n2 == /dev/nvme* ]] 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@526 -- # echo /dev/nvme0n2 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@524 -- # read -r dev _ 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@525 -- # [[ /dev/nvme0n1 == /dev/nvme* ]] 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@526 -- # echo /dev/nvme0n1 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@524 -- # read -r dev _ 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@35 -- # [[ -z /dev/nvme0n2 00:10:59.591 /dev/nvme0n1 ]] 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@59 -- # devs=($(get_nvme_devs)) 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@59 -- # get_nvme_devs 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@522 -- # local dev _ 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@524 -- # read -r dev _ 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@521 -- # nvme list 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@525 -- # [[ Node == /dev/nvme* ]] 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@524 -- # read -r dev _ 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@525 -- # [[ --------------------- == /dev/nvme* ]] 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@524 -- # read -r dev _ 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@525 -- # [[ /dev/nvme0n2 == /dev/nvme* ]] 00:10:59.591 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@526 -- # echo /dev/nvme0n2 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@524 -- # read -r dev _ 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@525 -- # [[ /dev/nvme0n1 == /dev/nvme* ]] 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@526 -- # echo /dev/nvme0n1 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@524 -- # read -r dev _ 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@59 -- # nvme_num=2 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@60 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:10:59.592 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@61 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@1219 -- # local i=0 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@1231 -- # return 0 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@62 -- # (( nvme_num <= nvme_num_before_connection )) 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@67 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@559 -- # xtrace_disable 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@68 -- # trap - SIGINT SIGTERM EXIT 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- target/nvme_cli.sh@70 -- # nvmftestfini 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@488 -- # nvmfcleanup 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@117 -- # sync 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@120 -- # set +e 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@121 -- # for i in {1..20} 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:10:59.592 rmmod nvme_tcp 00:10:59.592 rmmod nvme_fabrics 00:10:59.592 rmmod nvme_keyring 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@124 -- # set -e 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@125 -- # return 0 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@489 -- # '[' -n 507992 ']' 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@490 -- # killprocess 507992 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@948 -- # '[' -z 507992 ']' 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@952 -- # kill -0 507992 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@953 -- # uname 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 507992 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@966 -- # echo 'killing process with pid 507992' 00:10:59.592 killing process with pid 507992 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@967 -- # kill 507992 00:10:59.592 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@972 -- # wait 507992 00:10:59.851 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:10:59.851 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:10:59.851 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:10:59.851 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:10:59.851 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@278 -- # remove_spdk_ns 00:10:59.851 17:56:53 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:10:59.851 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:10:59.851 17:56:53 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:11:02.390 17:56:55 nvmf_tcp.nvmf_nvme_cli -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:11:02.390 00:11:02.390 real 0m12.215s 00:11:02.390 user 0m20.059s 00:11:02.390 sys 0m4.472s 00:11:02.390 17:56:55 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@1124 -- # xtrace_disable 00:11:02.390 17:56:55 nvmf_tcp.nvmf_nvme_cli -- common/autotest_common.sh@10 -- # set +x 00:11:02.390 ************************************ 00:11:02.390 END TEST nvmf_nvme_cli 00:11:02.390 ************************************ 00:11:02.391 17:56:55 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:11:02.391 17:56:55 nvmf_tcp -- nvmf/nvmf.sh@40 -- # [[ 1 -eq 1 ]] 00:11:02.391 17:56:55 nvmf_tcp -- nvmf/nvmf.sh@41 -- # run_test nvmf_vfio_user /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_vfio_user.sh --transport=tcp 00:11:02.391 17:56:55 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:11:02.391 17:56:55 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:11:02.391 17:56:55 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:11:02.391 ************************************ 00:11:02.391 START TEST nvmf_vfio_user 00:11:02.391 ************************************ 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_vfio_user.sh --transport=tcp 00:11:02.391 * Looking for test storage... 00:11:02.391 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@7 -- # uname -s 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- paths/export.sh@5 -- # export PATH 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@47 -- # : 0 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- nvmf/common.sh@51 -- # have_pci_nics=0 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@12 -- # MALLOC_BDEV_SIZE=64 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@14 -- # NUM_DEVICES=2 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@16 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@18 -- # export TEST_TRANSPORT=VFIOUSER 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@18 -- # TEST_TRANSPORT=VFIOUSER 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@47 -- # rm -rf /var/run/vfio-user 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@103 -- # setup_nvmf_vfio_user '' '' 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@51 -- # local nvmf_app_args= 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@52 -- # local transport_args= 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@55 -- # nvmfpid=509286 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@57 -- # echo 'Process pid: 509286' 00:11:02.391 Process pid: 509286 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@59 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@60 -- # waitforlisten 509286 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m '[0,1,2,3]' 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@829 -- # '[' -z 509286 ']' 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@834 -- # local max_retries=100 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:02.391 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@838 -- # xtrace_disable 00:11:02.391 17:56:55 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@10 -- # set +x 00:11:02.391 [2024-07-15 17:56:55.827597] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:11:02.391 [2024-07-15 17:56:55.827644] [ DPDK EAL parameters: nvmf -l 0,1,2,3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:02.391 EAL: No free 2048 kB hugepages reported on node 1 00:11:02.391 [2024-07-15 17:56:55.882119] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:11:02.391 [2024-07-15 17:56:55.963182] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:02.391 [2024-07-15 17:56:55.963217] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:02.391 [2024-07-15 17:56:55.963227] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:02.391 [2024-07-15 17:56:55.963233] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:02.391 [2024-07-15 17:56:55.963239] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:02.391 [2024-07-15 17:56:55.963281] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:11:02.391 [2024-07-15 17:56:55.963300] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:11:02.391 [2024-07-15 17:56:55.963387] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:11:02.391 [2024-07-15 17:56:55.963388] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:11:02.961 17:56:56 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:11:02.961 17:56:56 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@862 -- # return 0 00:11:02.961 17:56:56 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@62 -- # sleep 1 00:11:04.338 17:56:57 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t VFIOUSER 00:11:04.338 17:56:57 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@66 -- # mkdir -p /var/run/vfio-user 00:11:04.338 17:56:57 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # seq 1 2 00:11:04.338 17:56:57 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:11:04.338 17:56:57 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user1/1 00:11:04.338 17:56:57 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:11:04.338 Malloc1 00:11:04.338 17:56:58 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode1 -a -s SPDK1 00:11:04.597 17:56:58 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode1 Malloc1 00:11:04.857 17:56:58 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode1 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user1/1 -s 0 00:11:05.116 17:56:58 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:11:05.116 17:56:58 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user2/2 00:11:05.116 17:56:58 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc2 00:11:05.116 Malloc2 00:11:05.116 17:56:58 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode2 -a -s SPDK2 00:11:05.375 17:56:58 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode2 Malloc2 00:11:05.634 17:56:59 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode2 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user2/2 -s 0 00:11:05.634 17:56:59 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@104 -- # run_nvmf_vfio_user 00:11:05.634 17:56:59 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@80 -- # seq 1 2 00:11:05.634 17:56:59 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@80 -- # for i in $(seq 1 $NUM_DEVICES) 00:11:05.634 17:56:59 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@81 -- # test_traddr=/var/run/vfio-user/domain/vfio-user1/1 00:11:05.634 17:56:59 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@82 -- # test_subnqn=nqn.2019-07.io.spdk:cnode1 00:11:05.634 17:56:59 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -g -L nvme -L nvme_vfio -L vfio_pci 00:11:05.634 [2024-07-15 17:56:59.349618] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:11:05.634 [2024-07-15 17:56:59.349649] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --single-file-segments --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid509793 ] 00:11:05.634 EAL: No free 2048 kB hugepages reported on node 1 00:11:05.896 [2024-07-15 17:56:59.379761] nvme_vfio_user.c: 259:nvme_vfio_ctrlr_scan: *DEBUG*: Scan controller : /var/run/vfio-user/domain/vfio-user1/1 00:11:05.896 [2024-07-15 17:56:59.387558] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 0, Size 0x2000, Offset 0x0, Flags 0xf, Cap offset 32 00:11:05.896 [2024-07-15 17:56:59.387577] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0x1000, Offset 0x1000, Map addr 0x7fa82e29c000 00:11:05.896 [2024-07-15 17:56:59.388560] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 1, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:11:05.896 [2024-07-15 17:56:59.389555] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 2, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:11:05.896 [2024-07-15 17:56:59.390561] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 3, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:11:05.896 [2024-07-15 17:56:59.391565] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 4, Size 0x2000, Offset 0x0, Flags 0x3, Cap offset 0 00:11:05.896 [2024-07-15 17:56:59.392572] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 5, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:11:05.896 [2024-07-15 17:56:59.393577] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 6, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:11:05.896 [2024-07-15 17:56:59.394577] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 7, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:11:05.896 [2024-07-15 17:56:59.395583] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 8, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:11:05.896 [2024-07-15 17:56:59.396592] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 9, Size 0xc000, Offset 0x0, Flags 0xf, Cap offset 32 00:11:05.896 [2024-07-15 17:56:59.396601] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0xb000, Offset 0x1000, Map addr 0x7fa82e291000 00:11:05.896 [2024-07-15 17:56:59.397546] vfio_user_pci.c: 65:vfio_add_mr: *DEBUG*: Add memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:11:05.896 [2024-07-15 17:56:59.408155] vfio_user_pci.c: 386:spdk_vfio_user_setup: *DEBUG*: Device vfio-user0, Path /var/run/vfio-user/domain/vfio-user1/1/cntrl Setup Successfully 00:11:05.896 [2024-07-15 17:56:59.408175] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to connect adminq (no timeout) 00:11:05.896 [2024-07-15 17:56:59.412686] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x0, value 0x201e0100ff 00:11:05.896 [2024-07-15 17:56:59.412724] nvme_pcie_common.c: 132:nvme_pcie_qpair_construct: *INFO*: max_completions_cap = 64 num_trackers = 192 00:11:05.896 [2024-07-15 17:56:59.412796] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for connect adminq (no timeout) 00:11:05.896 [2024-07-15 17:56:59.412813] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to read vs (no timeout) 00:11:05.896 [2024-07-15 17:56:59.412818] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to read vs wait for vs (no timeout) 00:11:05.896 [2024-07-15 17:56:59.413685] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x8, value 0x10300 00:11:05.896 [2024-07-15 17:56:59.413695] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to read cap (no timeout) 00:11:05.896 [2024-07-15 17:56:59.413701] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to read cap wait for cap (no timeout) 00:11:05.896 [2024-07-15 17:56:59.414689] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x0, value 0x201e0100ff 00:11:05.896 [2024-07-15 17:56:59.414697] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to check en (no timeout) 00:11:05.896 [2024-07-15 17:56:59.414703] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to check en wait for cc (timeout 15000 ms) 00:11:05.896 [2024-07-15 17:56:59.415692] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x0 00:11:05.896 [2024-07-15 17:56:59.415700] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:11:05.896 [2024-07-15 17:56:59.416697] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x1c, value 0x0 00:11:05.896 [2024-07-15 17:56:59.416706] nvme_ctrlr.c:3869:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] CC.EN = 0 && CSTS.RDY = 0 00:11:05.896 [2024-07-15 17:56:59.416710] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to controller is disabled (timeout 15000 ms) 00:11:05.896 [2024-07-15 17:56:59.416716] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:11:05.896 [2024-07-15 17:56:59.416821] nvme_ctrlr.c:4062:nvme_ctrlr_process_init: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] Setting CC.EN = 1 00:11:05.896 [2024-07-15 17:56:59.416825] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:11:05.896 [2024-07-15 17:56:59.416829] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x28, value 0x2000003c0000 00:11:05.896 [2024-07-15 17:56:59.417712] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x30, value 0x2000003be000 00:11:05.896 [2024-07-15 17:56:59.418715] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x24, value 0xff00ff 00:11:05.896 [2024-07-15 17:56:59.419719] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x460001 00:11:05.896 [2024-07-15 17:56:59.420719] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:11:05.896 [2024-07-15 17:56:59.420780] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:11:05.896 [2024-07-15 17:56:59.421731] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x1c, value 0x1 00:11:05.896 [2024-07-15 17:56:59.421741] nvme_ctrlr.c:3904:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:11:05.896 [2024-07-15 17:56:59.421746] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to reset admin queue (timeout 30000 ms) 00:11:05.896 [2024-07-15 17:56:59.421763] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to identify controller (no timeout) 00:11:05.896 [2024-07-15 17:56:59.421772] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for identify controller (timeout 30000 ms) 00:11:05.896 [2024-07-15 17:56:59.421786] nvme_pcie_common.c:1201:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:11:05.896 [2024-07-15 17:56:59.421791] nvme_pcie_common.c:1229:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:11:05.896 [2024-07-15 17:56:59.421804] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000001 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:11:05.896 [2024-07-15 17:56:59.421842] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0001 p:1 m:0 dnr:0 00:11:05.896 [2024-07-15 17:56:59.421851] nvme_ctrlr.c:2053:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] transport max_xfer_size 131072 00:11:05.896 [2024-07-15 17:56:59.421859] nvme_ctrlr.c:2057:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] MDTS max_xfer_size 131072 00:11:05.896 [2024-07-15 17:56:59.421863] nvme_ctrlr.c:2060:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] CNTLID 0x0001 00:11:05.896 [2024-07-15 17:56:59.421867] nvme_ctrlr.c:2071:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] Identify CNTLID 0x0001 != Connect CNTLID 0x0000 00:11:05.896 [2024-07-15 17:56:59.421871] nvme_ctrlr.c:2084:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] transport max_sges 1 00:11:05.896 [2024-07-15 17:56:59.421875] nvme_ctrlr.c:2099:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] fuses compare and write: 1 00:11:05.896 [2024-07-15 17:56:59.421879] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to configure AER (timeout 30000 ms) 00:11:05.896 [2024-07-15 17:56:59.421886] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for configure aer (timeout 30000 ms) 00:11:05.896 [2024-07-15 17:56:59.421895] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:191 cdw10:0000000b PRP1 0x0 PRP2 0x0 00:11:05.896 [2024-07-15 17:56:59.421905] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0002 p:1 m:0 dnr:0 00:11:05.896 [2024-07-15 17:56:59.421916] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:190 nsid:0 cdw10:00000000 cdw11:00000000 00:11:05.896 [2024-07-15 17:56:59.421924] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:189 nsid:0 cdw10:00000000 cdw11:00000000 00:11:05.896 [2024-07-15 17:56:59.421931] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:188 nsid:0 cdw10:00000000 cdw11:00000000 00:11:05.896 [2024-07-15 17:56:59.421939] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:187 nsid:0 cdw10:00000000 cdw11:00000000 00:11:05.896 [2024-07-15 17:56:59.421943] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to set keep alive timeout (timeout 30000 ms) 00:11:05.896 [2024-07-15 17:56:59.421950] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:11:05.896 [2024-07-15 17:56:59.421958] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:191 cdw10:0000000f PRP1 0x0 PRP2 0x0 00:11:05.896 [2024-07-15 17:56:59.421966] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0007 p:1 m:0 dnr:0 00:11:05.896 [2024-07-15 17:56:59.421973] nvme_ctrlr.c:3010:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] Controller adjusted keep alive timeout to 0 ms 00:11:05.897 [2024-07-15 17:56:59.421977] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to identify controller iocs specific (timeout 30000 ms) 00:11:05.897 [2024-07-15 17:56:59.421982] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to set number of queues (timeout 30000 ms) 00:11:05.897 [2024-07-15 17:56:59.421988] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for set number of queues (timeout 30000 ms) 00:11:05.897 [2024-07-15 17:56:59.421995] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:11:05.897 [2024-07-15 17:56:59.422004] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:0008 p:1 m:0 dnr:0 00:11:05.897 [2024-07-15 17:56:59.422052] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to identify active ns (timeout 30000 ms) 00:11:05.897 [2024-07-15 17:56:59.422059] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for identify active ns (timeout 30000 ms) 00:11:05.897 [2024-07-15 17:56:59.422065] nvme_pcie_common.c:1201:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f9000 len:4096 00:11:05.897 [2024-07-15 17:56:59.422069] nvme_pcie_common.c:1229:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f9000 00:11:05.897 [2024-07-15 17:56:59.422075] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000002 cdw11:00000000 PRP1 0x2000002f9000 PRP2 0x0 00:11:05.897 [2024-07-15 17:56:59.422091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0009 p:1 m:0 dnr:0 00:11:05.897 [2024-07-15 17:56:59.422099] nvme_ctrlr.c:4693:spdk_nvme_ctrlr_get_ns: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] Namespace 1 was added 00:11:05.897 [2024-07-15 17:56:59.422107] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to identify ns (timeout 30000 ms) 00:11:05.897 [2024-07-15 17:56:59.422113] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for identify ns (timeout 30000 ms) 00:11:05.897 [2024-07-15 17:56:59.422119] nvme_pcie_common.c:1201:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:11:05.897 [2024-07-15 17:56:59.422123] nvme_pcie_common.c:1229:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:11:05.897 [2024-07-15 17:56:59.422128] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000000 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:11:05.897 [2024-07-15 17:56:59.422141] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000a p:1 m:0 dnr:0 00:11:05.897 [2024-07-15 17:56:59.422151] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to identify namespace id descriptors (timeout 30000 ms) 00:11:05.897 [2024-07-15 17:56:59.422158] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to wait for identify namespace id descriptors (timeout 30000 ms) 00:11:05.897 [2024-07-15 17:56:59.422164] nvme_pcie_common.c:1201:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:11:05.897 [2024-07-15 17:56:59.422168] nvme_pcie_common.c:1229:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:11:05.897 [2024-07-15 17:56:59.422173] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:11:05.897 [2024-07-15 17:56:59.422183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000b p:1 m:0 dnr:0 00:11:05.897 [2024-07-15 17:56:59.422191] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to identify ns iocs specific (timeout 30000 ms) 00:11:05.897 [2024-07-15 17:56:59.422196] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to set supported log pages (timeout 30000 ms) 00:11:05.897 [2024-07-15 17:56:59.422203] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to set supported features (timeout 30000 ms) 00:11:05.897 [2024-07-15 17:56:59.422208] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to set host behavior support feature (timeout 30000 ms) 00:11:05.897 [2024-07-15 17:56:59.422212] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to set doorbell buffer config (timeout 30000 ms) 00:11:05.897 [2024-07-15 17:56:59.422217] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to set host ID (timeout 30000 ms) 00:11:05.897 [2024-07-15 17:56:59.422221] nvme_ctrlr.c:3110:nvme_ctrlr_set_host_id: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] NVMe-oF transport - not sending Set Features - Host ID 00:11:05.897 [2024-07-15 17:56:59.422229] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to transport ready (timeout 30000 ms) 00:11:05.897 [2024-07-15 17:56:59.422234] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] setting state to ready (no timeout) 00:11:05.897 [2024-07-15 17:56:59.422249] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ARBITRATION cid:191 cdw10:00000001 PRP1 0x0 PRP2 0x0 00:11:05.897 [2024-07-15 17:56:59.422262] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000c p:1 m:0 dnr:0 00:11:05.897 [2024-07-15 17:56:59.422272] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES POWER MANAGEMENT cid:191 cdw10:00000002 PRP1 0x0 PRP2 0x0 00:11:05.897 [2024-07-15 17:56:59.422282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000d p:1 m:0 dnr:0 00:11:05.897 [2024-07-15 17:56:59.422291] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES TEMPERATURE THRESHOLD cid:191 cdw10:00000004 PRP1 0x0 PRP2 0x0 00:11:05.897 [2024-07-15 17:56:59.422303] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000e p:1 m:0 dnr:0 00:11:05.897 [2024-07-15 17:56:59.422312] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:11:05.897 [2024-07-15 17:56:59.422322] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:000f p:1 m:0 dnr:0 00:11:05.897 [2024-07-15 17:56:59.422334] nvme_pcie_common.c:1201:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f6000 len:8192 00:11:05.897 [2024-07-15 17:56:59.422338] nvme_pcie_common.c:1229:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f6000 00:11:05.897 [2024-07-15 17:56:59.422341] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp[0] = 0x2000002f7000 00:11:05.897 [2024-07-15 17:56:59.422344] nvme_pcie_common.c:1254:nvme_pcie_prp_list_append: *DEBUG*: prp2 = 0x2000002f7000 00:11:05.897 [2024-07-15 17:56:59.422349] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:191 nsid:ffffffff cdw10:07ff0001 cdw11:00000000 PRP1 0x2000002f6000 PRP2 0x2000002f7000 00:11:05.897 [2024-07-15 17:56:59.422356] nvme_pcie_common.c:1201:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fc000 len:512 00:11:05.897 [2024-07-15 17:56:59.422360] nvme_pcie_common.c:1229:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fc000 00:11:05.897 [2024-07-15 17:56:59.422365] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:186 nsid:ffffffff cdw10:007f0002 cdw11:00000000 PRP1 0x2000002fc000 PRP2 0x0 00:11:05.897 [2024-07-15 17:56:59.422371] nvme_pcie_common.c:1201:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:512 00:11:05.897 [2024-07-15 17:56:59.422376] nvme_pcie_common.c:1229:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:11:05.897 [2024-07-15 17:56:59.422382] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:185 nsid:ffffffff cdw10:007f0003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:11:05.897 [2024-07-15 17:56:59.422388] nvme_pcie_common.c:1201:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f4000 len:4096 00:11:05.897 [2024-07-15 17:56:59.422392] nvme_pcie_common.c:1229:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f4000 00:11:05.897 [2024-07-15 17:56:59.422398] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:184 nsid:ffffffff cdw10:03ff0005 cdw11:00000000 PRP1 0x2000002f4000 PRP2 0x0 00:11:05.897 [2024-07-15 17:56:59.422404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0010 p:1 m:0 dnr:0 00:11:05.897 [2024-07-15 17:56:59.422415] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:186 cdw0:0 sqhd:0011 p:1 m:0 dnr:0 00:11:05.897 [2024-07-15 17:56:59.422425] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:185 cdw0:0 sqhd:0012 p:1 m:0 dnr:0 00:11:05.897 [2024-07-15 17:56:59.422431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0013 p:1 m:0 dnr:0 00:11:05.897 ===================================================== 00:11:05.897 NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:11:05.897 ===================================================== 00:11:05.897 Controller Capabilities/Features 00:11:05.897 ================================ 00:11:05.897 Vendor ID: 4e58 00:11:05.897 Subsystem Vendor ID: 4e58 00:11:05.897 Serial Number: SPDK1 00:11:05.897 Model Number: SPDK bdev Controller 00:11:05.897 Firmware Version: 24.09 00:11:05.897 Recommended Arb Burst: 6 00:11:05.897 IEEE OUI Identifier: 8d 6b 50 00:11:05.897 Multi-path I/O 00:11:05.897 May have multiple subsystem ports: Yes 00:11:05.897 May have multiple controllers: Yes 00:11:05.897 Associated with SR-IOV VF: No 00:11:05.897 Max Data Transfer Size: 131072 00:11:05.897 Max Number of Namespaces: 32 00:11:05.897 Max Number of I/O Queues: 127 00:11:05.897 NVMe Specification Version (VS): 1.3 00:11:05.897 NVMe Specification Version (Identify): 1.3 00:11:05.897 Maximum Queue Entries: 256 00:11:05.897 Contiguous Queues Required: Yes 00:11:05.897 Arbitration Mechanisms Supported 00:11:05.897 Weighted Round Robin: Not Supported 00:11:05.897 Vendor Specific: Not Supported 00:11:05.897 Reset Timeout: 15000 ms 00:11:05.897 Doorbell Stride: 4 bytes 00:11:05.897 NVM Subsystem Reset: Not Supported 00:11:05.897 Command Sets Supported 00:11:05.897 NVM Command Set: Supported 00:11:05.897 Boot Partition: Not Supported 00:11:05.897 Memory Page Size Minimum: 4096 bytes 00:11:05.897 Memory Page Size Maximum: 4096 bytes 00:11:05.897 Persistent Memory Region: Not Supported 00:11:05.897 Optional Asynchronous Events Supported 00:11:05.897 Namespace Attribute Notices: Supported 00:11:05.897 Firmware Activation Notices: Not Supported 00:11:05.897 ANA Change Notices: Not Supported 00:11:05.897 PLE Aggregate Log Change Notices: Not Supported 00:11:05.897 LBA Status Info Alert Notices: Not Supported 00:11:05.897 EGE Aggregate Log Change Notices: Not Supported 00:11:05.897 Normal NVM Subsystem Shutdown event: Not Supported 00:11:05.897 Zone Descriptor Change Notices: Not Supported 00:11:05.897 Discovery Log Change Notices: Not Supported 00:11:05.897 Controller Attributes 00:11:05.897 128-bit Host Identifier: Supported 00:11:05.897 Non-Operational Permissive Mode: Not Supported 00:11:05.897 NVM Sets: Not Supported 00:11:05.897 Read Recovery Levels: Not Supported 00:11:05.897 Endurance Groups: Not Supported 00:11:05.897 Predictable Latency Mode: Not Supported 00:11:05.897 Traffic Based Keep ALive: Not Supported 00:11:05.897 Namespace Granularity: Not Supported 00:11:05.897 SQ Associations: Not Supported 00:11:05.897 UUID List: Not Supported 00:11:05.897 Multi-Domain Subsystem: Not Supported 00:11:05.897 Fixed Capacity Management: Not Supported 00:11:05.897 Variable Capacity Management: Not Supported 00:11:05.897 Delete Endurance Group: Not Supported 00:11:05.897 Delete NVM Set: Not Supported 00:11:05.897 Extended LBA Formats Supported: Not Supported 00:11:05.897 Flexible Data Placement Supported: Not Supported 00:11:05.897 00:11:05.897 Controller Memory Buffer Support 00:11:05.897 ================================ 00:11:05.898 Supported: No 00:11:05.898 00:11:05.898 Persistent Memory Region Support 00:11:05.898 ================================ 00:11:05.898 Supported: No 00:11:05.898 00:11:05.898 Admin Command Set Attributes 00:11:05.898 ============================ 00:11:05.898 Security Send/Receive: Not Supported 00:11:05.898 Format NVM: Not Supported 00:11:05.898 Firmware Activate/Download: Not Supported 00:11:05.898 Namespace Management: Not Supported 00:11:05.898 Device Self-Test: Not Supported 00:11:05.898 Directives: Not Supported 00:11:05.898 NVMe-MI: Not Supported 00:11:05.898 Virtualization Management: Not Supported 00:11:05.898 Doorbell Buffer Config: Not Supported 00:11:05.898 Get LBA Status Capability: Not Supported 00:11:05.898 Command & Feature Lockdown Capability: Not Supported 00:11:05.898 Abort Command Limit: 4 00:11:05.898 Async Event Request Limit: 4 00:11:05.898 Number of Firmware Slots: N/A 00:11:05.898 Firmware Slot 1 Read-Only: N/A 00:11:05.898 Firmware Activation Without Reset: N/A 00:11:05.898 Multiple Update Detection Support: N/A 00:11:05.898 Firmware Update Granularity: No Information Provided 00:11:05.898 Per-Namespace SMART Log: No 00:11:05.898 Asymmetric Namespace Access Log Page: Not Supported 00:11:05.898 Subsystem NQN: nqn.2019-07.io.spdk:cnode1 00:11:05.898 Command Effects Log Page: Supported 00:11:05.898 Get Log Page Extended Data: Supported 00:11:05.898 Telemetry Log Pages: Not Supported 00:11:05.898 Persistent Event Log Pages: Not Supported 00:11:05.898 Supported Log Pages Log Page: May Support 00:11:05.898 Commands Supported & Effects Log Page: Not Supported 00:11:05.898 Feature Identifiers & Effects Log Page:May Support 00:11:05.898 NVMe-MI Commands & Effects Log Page: May Support 00:11:05.898 Data Area 4 for Telemetry Log: Not Supported 00:11:05.898 Error Log Page Entries Supported: 128 00:11:05.898 Keep Alive: Supported 00:11:05.898 Keep Alive Granularity: 10000 ms 00:11:05.898 00:11:05.898 NVM Command Set Attributes 00:11:05.898 ========================== 00:11:05.898 Submission Queue Entry Size 00:11:05.898 Max: 64 00:11:05.898 Min: 64 00:11:05.898 Completion Queue Entry Size 00:11:05.898 Max: 16 00:11:05.898 Min: 16 00:11:05.898 Number of Namespaces: 32 00:11:05.898 Compare Command: Supported 00:11:05.898 Write Uncorrectable Command: Not Supported 00:11:05.898 Dataset Management Command: Supported 00:11:05.898 Write Zeroes Command: Supported 00:11:05.898 Set Features Save Field: Not Supported 00:11:05.898 Reservations: Not Supported 00:11:05.898 Timestamp: Not Supported 00:11:05.898 Copy: Supported 00:11:05.898 Volatile Write Cache: Present 00:11:05.898 Atomic Write Unit (Normal): 1 00:11:05.898 Atomic Write Unit (PFail): 1 00:11:05.898 Atomic Compare & Write Unit: 1 00:11:05.898 Fused Compare & Write: Supported 00:11:05.898 Scatter-Gather List 00:11:05.898 SGL Command Set: Supported (Dword aligned) 00:11:05.898 SGL Keyed: Not Supported 00:11:05.898 SGL Bit Bucket Descriptor: Not Supported 00:11:05.898 SGL Metadata Pointer: Not Supported 00:11:05.898 Oversized SGL: Not Supported 00:11:05.898 SGL Metadata Address: Not Supported 00:11:05.898 SGL Offset: Not Supported 00:11:05.898 Transport SGL Data Block: Not Supported 00:11:05.898 Replay Protected Memory Block: Not Supported 00:11:05.898 00:11:05.898 Firmware Slot Information 00:11:05.898 ========================= 00:11:05.898 Active slot: 1 00:11:05.898 Slot 1 Firmware Revision: 24.09 00:11:05.898 00:11:05.898 00:11:05.898 Commands Supported and Effects 00:11:05.898 ============================== 00:11:05.898 Admin Commands 00:11:05.898 -------------- 00:11:05.898 Get Log Page (02h): Supported 00:11:05.898 Identify (06h): Supported 00:11:05.898 Abort (08h): Supported 00:11:05.898 Set Features (09h): Supported 00:11:05.898 Get Features (0Ah): Supported 00:11:05.898 Asynchronous Event Request (0Ch): Supported 00:11:05.898 Keep Alive (18h): Supported 00:11:05.898 I/O Commands 00:11:05.898 ------------ 00:11:05.898 Flush (00h): Supported LBA-Change 00:11:05.898 Write (01h): Supported LBA-Change 00:11:05.898 Read (02h): Supported 00:11:05.898 Compare (05h): Supported 00:11:05.898 Write Zeroes (08h): Supported LBA-Change 00:11:05.898 Dataset Management (09h): Supported LBA-Change 00:11:05.898 Copy (19h): Supported LBA-Change 00:11:05.898 00:11:05.898 Error Log 00:11:05.898 ========= 00:11:05.898 00:11:05.898 Arbitration 00:11:05.898 =========== 00:11:05.898 Arbitration Burst: 1 00:11:05.898 00:11:05.898 Power Management 00:11:05.898 ================ 00:11:05.898 Number of Power States: 1 00:11:05.898 Current Power State: Power State #0 00:11:05.898 Power State #0: 00:11:05.898 Max Power: 0.00 W 00:11:05.898 Non-Operational State: Operational 00:11:05.898 Entry Latency: Not Reported 00:11:05.898 Exit Latency: Not Reported 00:11:05.898 Relative Read Throughput: 0 00:11:05.898 Relative Read Latency: 0 00:11:05.898 Relative Write Throughput: 0 00:11:05.898 Relative Write Latency: 0 00:11:05.898 Idle Power: Not Reported 00:11:05.898 Active Power: Not Reported 00:11:05.898 Non-Operational Permissive Mode: Not Supported 00:11:05.898 00:11:05.898 Health Information 00:11:05.898 ================== 00:11:05.898 Critical Warnings: 00:11:05.898 Available Spare Space: OK 00:11:05.898 Temperature: OK 00:11:05.898 Device Reliability: OK 00:11:05.898 Read Only: No 00:11:05.898 Volatile Memory Backup: OK 00:11:05.898 Current Temperature: 0 Kelvin (-273 Celsius) 00:11:05.898 Temperature Threshold: 0 Kelvin (-273 Celsius) 00:11:05.898 Available Spare: 0% 00:11:05.898 Available Sp[2024-07-15 17:56:59.422520] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ERROR_RECOVERY cid:184 cdw10:00000005 PRP1 0x0 PRP2 0x0 00:11:05.898 [2024-07-15 17:56:59.422533] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0014 p:1 m:0 dnr:0 00:11:05.898 [2024-07-15 17:56:59.422560] nvme_ctrlr.c:4357:nvme_ctrlr_destruct_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] Prepare to destruct SSD 00:11:05.898 [2024-07-15 17:56:59.422568] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:190 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:05.898 [2024-07-15 17:56:59.422574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:189 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:05.898 [2024-07-15 17:56:59.422579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:188 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:05.898 [2024-07-15 17:56:59.422584] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:187 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:05.898 [2024-07-15 17:56:59.422738] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x460001 00:11:05.898 [2024-07-15 17:56:59.422747] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x14, value 0x464001 00:11:05.898 [2024-07-15 17:56:59.423747] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:11:05.898 [2024-07-15 17:56:59.423793] nvme_ctrlr.c:1147:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] RTD3E = 0 us 00:11:05.898 [2024-07-15 17:56:59.423799] nvme_ctrlr.c:1150:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] shutdown timeout = 10000 ms 00:11:05.898 [2024-07-15 17:56:59.424753] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user1/1: offset 0x1c, value 0x9 00:11:05.898 [2024-07-15 17:56:59.424764] nvme_ctrlr.c:1269:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user1/1] shutdown complete in 0 milliseconds 00:11:05.898 [2024-07-15 17:56:59.424813] vfio_user_pci.c: 399:spdk_vfio_user_release: *DEBUG*: Release file /var/run/vfio-user/domain/vfio-user1/1/cntrl 00:11:05.898 [2024-07-15 17:56:59.430232] vfio_user_pci.c: 96:vfio_remove_mr: *DEBUG*: Remove memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:11:05.898 are Threshold: 0% 00:11:05.898 Life Percentage Used: 0% 00:11:05.898 Data Units Read: 0 00:11:05.898 Data Units Written: 0 00:11:05.898 Host Read Commands: 0 00:11:05.898 Host Write Commands: 0 00:11:05.898 Controller Busy Time: 0 minutes 00:11:05.898 Power Cycles: 0 00:11:05.898 Power On Hours: 0 hours 00:11:05.898 Unsafe Shutdowns: 0 00:11:05.898 Unrecoverable Media Errors: 0 00:11:05.898 Lifetime Error Log Entries: 0 00:11:05.898 Warning Temperature Time: 0 minutes 00:11:05.898 Critical Temperature Time: 0 minutes 00:11:05.898 00:11:05.898 Number of Queues 00:11:05.898 ================ 00:11:05.898 Number of I/O Submission Queues: 127 00:11:05.898 Number of I/O Completion Queues: 127 00:11:05.898 00:11:05.898 Active Namespaces 00:11:05.898 ================= 00:11:05.898 Namespace ID:1 00:11:05.898 Error Recovery Timeout: Unlimited 00:11:05.898 Command Set Identifier: NVM (00h) 00:11:05.898 Deallocate: Supported 00:11:05.898 Deallocated/Unwritten Error: Not Supported 00:11:05.898 Deallocated Read Value: Unknown 00:11:05.898 Deallocate in Write Zeroes: Not Supported 00:11:05.898 Deallocated Guard Field: 0xFFFF 00:11:05.898 Flush: Supported 00:11:05.898 Reservation: Supported 00:11:05.898 Namespace Sharing Capabilities: Multiple Controllers 00:11:05.898 Size (in LBAs): 131072 (0GiB) 00:11:05.898 Capacity (in LBAs): 131072 (0GiB) 00:11:05.898 Utilization (in LBAs): 131072 (0GiB) 00:11:05.898 NGUID: D36E78CC2C2C4326A6A01BE2FFD8961C 00:11:05.898 UUID: d36e78cc-2c2c-4326-a6a0-1be2ffd8961c 00:11:05.899 Thin Provisioning: Not Supported 00:11:05.899 Per-NS Atomic Units: Yes 00:11:05.899 Atomic Boundary Size (Normal): 0 00:11:05.899 Atomic Boundary Size (PFail): 0 00:11:05.899 Atomic Boundary Offset: 0 00:11:05.899 Maximum Single Source Range Length: 65535 00:11:05.899 Maximum Copy Length: 65535 00:11:05.899 Maximum Source Range Count: 1 00:11:05.899 NGUID/EUI64 Never Reused: No 00:11:05.899 Namespace Write Protected: No 00:11:05.899 Number of LBA Formats: 1 00:11:05.899 Current LBA Format: LBA Format #00 00:11:05.899 LBA Format #00: Data Size: 512 Metadata Size: 0 00:11:05.899 00:11:05.899 17:56:59 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -s 256 -g -q 128 -o 4096 -w read -t 5 -c 0x2 00:11:05.899 EAL: No free 2048 kB hugepages reported on node 1 00:11:06.158 [2024-07-15 17:56:59.647053] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:11:11.431 Initializing NVMe Controllers 00:11:11.431 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:11:11.431 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 with lcore 1 00:11:11.431 Initialization complete. Launching workers. 00:11:11.431 ======================================================== 00:11:11.431 Latency(us) 00:11:11.431 Device Information : IOPS MiB/s Average min max 00:11:11.431 VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 from core 1: 39930.91 155.98 3205.12 964.83 6698.27 00:11:11.431 ======================================================== 00:11:11.431 Total : 39930.91 155.98 3205.12 964.83 6698.27 00:11:11.431 00:11:11.431 [2024-07-15 17:57:04.664997] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:11:11.431 17:57:04 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@85 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -s 256 -g -q 128 -o 4096 -w write -t 5 -c 0x2 00:11:11.431 EAL: No free 2048 kB hugepages reported on node 1 00:11:11.431 [2024-07-15 17:57:04.887015] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:11:16.700 Initializing NVMe Controllers 00:11:16.700 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:11:16.700 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 with lcore 1 00:11:16.700 Initialization complete. Launching workers. 00:11:16.700 ======================================================== 00:11:16.700 Latency(us) 00:11:16.700 Device Information : IOPS MiB/s Average min max 00:11:16.700 VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) NSID 1 from core 1: 16051.46 62.70 7979.67 4972.92 10979.75 00:11:16.700 ======================================================== 00:11:16.700 Total : 16051.46 62.70 7979.67 4972.92 10979.75 00:11:16.700 00:11:16.700 [2024-07-15 17:57:09.929312] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:11:16.700 17:57:09 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -g -q 32 -o 4096 -w randrw -M 50 -t 5 -c 0xE 00:11:16.700 EAL: No free 2048 kB hugepages reported on node 1 00:11:16.700 [2024-07-15 17:57:10.128266] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:11:22.039 [2024-07-15 17:57:15.196506] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:11:22.039 Initializing NVMe Controllers 00:11:22.039 Attaching to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:11:22.039 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user1/1:: nqn.2019-07.io.spdk:cnode1 00:11:22.039 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) with lcore 1 00:11:22.039 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) with lcore 2 00:11:22.039 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user1/1) with lcore 3 00:11:22.039 Initialization complete. Launching workers. 00:11:22.039 Starting thread on core 2 00:11:22.039 Starting thread on core 3 00:11:22.039 Starting thread on core 1 00:11:22.039 17:57:15 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -t 3 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -d 256 -g 00:11:22.039 EAL: No free 2048 kB hugepages reported on node 1 00:11:22.039 [2024-07-15 17:57:15.474620] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:11:25.429 [2024-07-15 17:57:18.676484] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:11:25.429 Initializing NVMe Controllers 00:11:25.429 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:11:25.429 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:11:25.429 Associating SPDK bdev Controller (SPDK1 ) with lcore 0 00:11:25.429 Associating SPDK bdev Controller (SPDK1 ) with lcore 1 00:11:25.429 Associating SPDK bdev Controller (SPDK1 ) with lcore 2 00:11:25.429 Associating SPDK bdev Controller (SPDK1 ) with lcore 3 00:11:25.429 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration run with configuration: 00:11:25.429 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -q 64 -s 131072 -w randrw -M 50 -l 0 -t 3 -c 0xf -m 0 -a 0 -b 0 -n 100000 -i -1 00:11:25.429 Initialization complete. Launching workers. 00:11:25.429 Starting thread on core 1 with urgent priority queue 00:11:25.429 Starting thread on core 2 with urgent priority queue 00:11:25.429 Starting thread on core 3 with urgent priority queue 00:11:25.429 Starting thread on core 0 with urgent priority queue 00:11:25.429 SPDK bdev Controller (SPDK1 ) core 0: 7447.00 IO/s 13.43 secs/100000 ios 00:11:25.430 SPDK bdev Controller (SPDK1 ) core 1: 5801.00 IO/s 17.24 secs/100000 ios 00:11:25.430 SPDK bdev Controller (SPDK1 ) core 2: 6321.67 IO/s 15.82 secs/100000 ios 00:11:25.430 SPDK bdev Controller (SPDK1 ) core 3: 5884.33 IO/s 16.99 secs/100000 ios 00:11:25.430 ======================================================== 00:11:25.430 00:11:25.430 17:57:18 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/hello_world -d 256 -g -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' 00:11:25.430 EAL: No free 2048 kB hugepages reported on node 1 00:11:25.430 [2024-07-15 17:57:18.951679] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:11:25.430 Initializing NVMe Controllers 00:11:25.430 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:11:25.430 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:11:25.430 Namespace ID: 1 size: 0GB 00:11:25.430 Initialization complete. 00:11:25.430 INFO: using host memory buffer for IO 00:11:25.430 Hello world! 00:11:25.430 [2024-07-15 17:57:18.985925] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:11:25.430 17:57:19 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/overhead/overhead -o 4096 -t 1 -H -g -d 256 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' 00:11:25.430 EAL: No free 2048 kB hugepages reported on node 1 00:11:25.688 [2024-07-15 17:57:19.260692] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:11:26.642 Initializing NVMe Controllers 00:11:26.642 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:11:26.642 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:11:26.642 Initialization complete. Launching workers. 00:11:26.642 submit (in ns) avg, min, max = 7599.1, 3219.1, 3999227.8 00:11:26.642 complete (in ns) avg, min, max = 20126.7, 1789.6, 4182721.7 00:11:26.642 00:11:26.642 Submit histogram 00:11:26.642 ================ 00:11:26.642 Range in us Cumulative Count 00:11:26.642 3.214 - 3.228: 0.0183% ( 3) 00:11:26.642 3.228 - 3.242: 0.0427% ( 4) 00:11:26.642 3.242 - 3.256: 0.0671% ( 4) 00:11:26.642 3.256 - 3.270: 0.1098% ( 7) 00:11:26.642 3.270 - 3.283: 0.6652% ( 91) 00:11:26.642 3.283 - 3.297: 3.3014% ( 432) 00:11:26.642 3.297 - 3.311: 8.1284% ( 791) 00:11:26.642 3.311 - 3.325: 13.6450% ( 904) 00:11:26.642 3.325 - 3.339: 19.6131% ( 978) 00:11:26.642 3.339 - 3.353: 26.1854% ( 1077) 00:11:26.642 3.353 - 3.367: 31.8911% ( 935) 00:11:26.642 3.367 - 3.381: 37.1575% ( 863) 00:11:26.642 3.381 - 3.395: 42.0028% ( 794) 00:11:26.642 3.395 - 3.409: 46.4148% ( 723) 00:11:26.642 3.409 - 3.423: 50.6072% ( 687) 00:11:26.642 3.423 - 3.437: 55.3976% ( 785) 00:11:26.642 3.437 - 3.450: 62.3604% ( 1141) 00:11:26.642 3.450 - 3.464: 67.4193% ( 829) 00:11:26.642 3.464 - 3.478: 72.0510% ( 759) 00:11:26.642 3.478 - 3.492: 78.1046% ( 992) 00:11:26.642 3.492 - 3.506: 81.7355% ( 595) 00:11:26.642 3.506 - 3.520: 84.1399% ( 394) 00:11:26.642 3.520 - 3.534: 85.4031% ( 207) 00:11:26.642 3.534 - 3.548: 86.0072% ( 99) 00:11:26.642 3.548 - 3.562: 86.4954% ( 80) 00:11:26.642 3.562 - 3.590: 87.3619% ( 142) 00:11:26.642 3.590 - 3.617: 88.8753% ( 248) 00:11:26.642 3.617 - 3.645: 90.6633% ( 293) 00:11:26.642 3.645 - 3.673: 92.4269% ( 289) 00:11:26.642 3.673 - 3.701: 94.1295% ( 279) 00:11:26.642 3.701 - 3.729: 95.8931% ( 289) 00:11:26.642 3.729 - 3.757: 97.2600% ( 224) 00:11:26.642 3.757 - 3.784: 98.0167% ( 124) 00:11:26.642 3.784 - 3.812: 98.5537% ( 88) 00:11:26.642 3.812 - 3.840: 98.9199% ( 60) 00:11:26.642 3.840 - 3.868: 99.1274% ( 34) 00:11:26.642 3.868 - 3.896: 99.2311% ( 17) 00:11:26.642 3.896 - 3.923: 99.2616% ( 5) 00:11:26.642 3.923 - 3.951: 99.2738% ( 2) 00:11:26.642 3.951 - 3.979: 99.2921% ( 3) 00:11:26.642 3.979 - 4.007: 99.3104% ( 3) 00:11:26.642 4.007 - 4.035: 99.3226% ( 2) 00:11:26.642 4.035 - 4.063: 99.3409% ( 3) 00:11:26.642 4.063 - 4.090: 99.3470% ( 1) 00:11:26.642 4.090 - 4.118: 99.3592% ( 2) 00:11:26.642 4.118 - 4.146: 99.3837% ( 4) 00:11:26.642 4.146 - 4.174: 99.3898% ( 1) 00:11:26.642 4.174 - 4.202: 99.4203% ( 5) 00:11:26.642 4.202 - 4.230: 99.4325% ( 2) 00:11:26.642 4.257 - 4.285: 99.4386% ( 1) 00:11:26.643 4.285 - 4.313: 99.4447% ( 1) 00:11:26.643 4.313 - 4.341: 99.4691% ( 4) 00:11:26.643 4.341 - 4.369: 99.4752% ( 1) 00:11:26.643 4.369 - 4.397: 99.4874% ( 2) 00:11:26.643 4.424 - 4.452: 99.4996% ( 2) 00:11:26.643 4.452 - 4.480: 99.5057% ( 1) 00:11:26.643 4.508 - 4.536: 99.5118% ( 1) 00:11:26.643 4.619 - 4.647: 99.5240% ( 2) 00:11:26.643 4.647 - 4.675: 99.5423% ( 3) 00:11:26.643 4.703 - 4.730: 99.5545% ( 2) 00:11:26.643 4.758 - 4.786: 99.5606% ( 1) 00:11:26.643 4.842 - 4.870: 99.5728% ( 2) 00:11:26.643 4.870 - 4.897: 99.5789% ( 1) 00:11:26.643 4.925 - 4.953: 99.5850% ( 1) 00:11:26.643 5.037 - 5.064: 99.5911% ( 1) 00:11:26.643 5.064 - 5.092: 99.5972% ( 1) 00:11:26.643 5.120 - 5.148: 99.6033% ( 1) 00:11:26.643 5.148 - 5.176: 99.6155% ( 2) 00:11:26.643 5.176 - 5.203: 99.6217% ( 1) 00:11:26.643 5.259 - 5.287: 99.6278% ( 1) 00:11:26.643 5.343 - 5.370: 99.6339% ( 1) 00:11:26.643 5.454 - 5.482: 99.6400% ( 1) 00:11:26.643 5.510 - 5.537: 99.6461% ( 1) 00:11:26.643 5.537 - 5.565: 99.6522% ( 1) 00:11:26.643 5.565 - 5.593: 99.6583% ( 1) 00:11:26.643 5.621 - 5.649: 99.6705% ( 2) 00:11:26.643 5.760 - 5.788: 99.6766% ( 1) 00:11:26.643 5.816 - 5.843: 99.6827% ( 1) 00:11:26.643 5.871 - 5.899: 99.6888% ( 1) 00:11:26.643 5.899 - 5.927: 99.6949% ( 1) 00:11:26.643 5.955 - 5.983: 99.7010% ( 1) 00:11:26.643 6.122 - 6.150: 99.7071% ( 1) 00:11:26.643 6.233 - 6.261: 99.7132% ( 1) 00:11:26.643 6.372 - 6.400: 99.7193% ( 1) 00:11:26.643 6.428 - 6.456: 99.7254% ( 1) 00:11:26.643 6.539 - 6.567: 99.7376% ( 2) 00:11:26.643 6.567 - 6.595: 99.7437% ( 1) 00:11:26.643 6.595 - 6.623: 99.7498% ( 1) 00:11:26.643 6.650 - 6.678: 99.7559% ( 1) 00:11:26.643 6.762 - 6.790: 99.7620% ( 1) 00:11:26.643 6.790 - 6.817: 99.7742% ( 2) 00:11:26.643 6.929 - 6.957: 99.7803% ( 1) 00:11:26.643 6.957 - 6.984: 99.7864% ( 1) 00:11:26.643 7.096 - 7.123: 99.7925% ( 1) 00:11:26.643 7.290 - 7.346: 99.7986% ( 1) 00:11:26.643 7.402 - 7.457: 99.8047% ( 1) 00:11:26.643 7.736 - 7.791: 99.8108% ( 1) 00:11:26.643 8.181 - 8.237: 99.8169% ( 1) 00:11:26.643 8.348 - 8.403: 99.8230% ( 1) 00:11:26.643 9.517 - 9.572: 99.8291% ( 1) 00:11:26.643 10.129 - 10.184: 99.8352% ( 1) 00:11:26.643 11.186 - 11.242: 99.8413% ( 1) 00:11:26.643 12.911 - 12.967: 99.8474% ( 1) 00:11:26.643 13.357 - 13.412: 99.8535% ( 1) 00:11:26.643 13.635 - 13.690: 99.8596% ( 1) 00:11:26.643 13.802 - 13.857: 99.8657% ( 1) 00:11:26.643 14.470 - 14.581: 99.8718% ( 1) 00:11:26.643 16.028 - 16.139: 99.8780% ( 1) 00:11:26.643 19.256 - 19.367: 99.8841% ( 1) 00:11:26.643 19.701 - 19.812: 99.8902% ( 1) 00:11:26.643 20.146 - 20.257: 99.8963% ( 1) 00:11:26.643 3989.148 - 4017.642: 100.0000% ( 17) 00:11:26.643 00:11:26.643 Complete histogram 00:11:26.643 ================== 00:11:26.643 Range in us Cumulative Count 00:11:26.643 1.781 - 1.795: 0.0122% ( 2) 00:11:26.643 1.795 - 1.809: 0.0183% ( 1) 00:11:26.643 1.809 - 1.823: 0.7018% ( 112) 00:11:26.643 1.823 - 1.837: 2.4288% ( 283) 00:11:26.643 1.837 - 1.850: 3.9238% ( 245) 00:11:26.643 1.850 - 1.864: 6.3160% ( 392) 00:11:26.643 1.864 - 1.878: 45.9328% ( 6492) 00:11:26.643 1.878 - 1.892: 86.4771% ( 6644) 00:11:26.643 1.892 - 1.906: 93.0067% ( 1070) 00:11:26.643 1.906 - 1.920: 95.0204% ( 330) 00:11:26.643 1.920 - 1.934: 95.6856% ( 109) 00:11:26.643 1.934 - 1.948: 96.6498% ( 158) 00:11:26.643 1.948 - 1.962: 97.8214% ( 192) 00:11:26.643 1.962 - 1.976: 98.4012% ( 95) 00:11:26.643 1.976 - 1.990: 98.5903% ( 31) 00:11:26.643 1.990 - 2.003: 98.6392% ( 8) 00:11:26.643 2.003 - 2.017: 98.6758% ( 6) 00:11:26.643 2.017 - 2.031: 98.7063% ( 5) 00:11:26.643 2.031 - 2.045: 98.7307% ( 4) 00:11:26.643 2.045 - 2.059: 98.7429% ( 2) 00:11:26.643 2.059 - 2.073: 98.8283% ( 14) 00:11:26.643 2.073 - 2.087: 98.9260% ( 16) 00:11:26.643 2.087 - 2.101: 98.9443% ( 3) 00:11:26.643 2.101 - 2.115: 98.9687% ( 4) 00:11:26.643 2.115 - 2.129: 98.9931% ( 4) 00:11:26.643 2.129 - 2.143: 98.9992% ( 1) 00:11:26.643 2.143 - 2.157: 99.0053% ( 1) 00:11:26.643 2.157 - 2.170: 99.0114% ( 1) 00:11:26.643 2.184 - 2.198: 99.0297% ( 3) 00:11:26.643 2.198 - 2.212: 99.0419% ( 2) 00:11:26.643 2.212 - 2.226: 99.0541% ( 2) 00:11:26.643 2.226 - 2.240: 99.0602% ( 1) 00:11:26.643 2.240 - 2.254: 99.0785% ( 3) 00:11:26.643 2.268 - 2.282: 99.0846% ( 1) 00:11:26.643 2.282 - 2.296: 99.0907% ( 1) 00:11:26.643 2.296 - 2.310: 99.1029% ( 2) 00:11:26.643 2.310 - 2.323: 99.1274% ( 4) 00:11:26.643 2.337 - 2.351: 99.1335% ( 1) 00:11:26.643 2.351 - 2.365: 99.1396% ( 1) 00:11:26.643 2.365 - 2.379: 99.1640% ( 4) 00:11:26.643 2.379 - 2.393: 99.1701% ( 1) 00:11:26.643 2.393 - 2.407: 99.1823% ( 2) 00:11:26.643 2.407 - 2.421: 99.1884% ( 1) 00:11:26.643 2.449 - 2.463: 99.2006% ( 2) 00:11:26.643 2.477 - 2.490: 99.2128% ( 2) 00:11:26.643 2.490 - 2.504: 99.2189% ( 1) 00:11:26.643 2.504 - 2.518: 99.2250% ( 1) 00:11:26.643 2.518 - 2.532: 99.2494% ( 4) 00:11:26.643 2.532 - 2.546: 99.2616% ( 2) 00:11:26.643 2.546 - 2.560: 99.2677% ( 1) 00:11:26.643 2.560 - 2.574: 99.2738% ( 1) 00:11:26.643 2.643 - 2.657: 99.2799% ( 1) 00:11:26.643 2.657 - 2.6[2024-07-15 17:57:20.282495] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:11:26.643 71: 99.2921% ( 2) 00:11:26.643 2.699 - 2.713: 99.2982% ( 1) 00:11:26.643 2.713 - 2.727: 99.3104% ( 2) 00:11:26.643 2.741 - 2.755: 99.3165% ( 1) 00:11:26.643 2.922 - 2.936: 99.3226% ( 1) 00:11:26.643 2.963 - 2.977: 99.3348% ( 2) 00:11:26.643 3.089 - 3.103: 99.3409% ( 1) 00:11:26.643 3.283 - 3.297: 99.3470% ( 1) 00:11:26.643 3.701 - 3.729: 99.3531% ( 1) 00:11:26.643 3.784 - 3.812: 99.3592% ( 1) 00:11:26.643 3.951 - 3.979: 99.3654% ( 1) 00:11:26.643 3.979 - 4.007: 99.3715% ( 1) 00:11:26.643 4.369 - 4.397: 99.3776% ( 1) 00:11:26.643 4.536 - 4.563: 99.3837% ( 1) 00:11:26.643 4.647 - 4.675: 99.4081% ( 4) 00:11:26.643 4.703 - 4.730: 99.4142% ( 1) 00:11:26.643 4.758 - 4.786: 99.4203% ( 1) 00:11:26.643 4.842 - 4.870: 99.4264% ( 1) 00:11:26.643 4.870 - 4.897: 99.4325% ( 1) 00:11:26.643 5.009 - 5.037: 99.4386% ( 1) 00:11:26.643 5.426 - 5.454: 99.4447% ( 1) 00:11:26.643 5.454 - 5.482: 99.4508% ( 1) 00:11:26.643 5.565 - 5.593: 99.4569% ( 1) 00:11:26.643 5.760 - 5.788: 99.4630% ( 1) 00:11:26.643 5.927 - 5.955: 99.4691% ( 1) 00:11:26.643 5.983 - 6.010: 99.4752% ( 1) 00:11:26.643 6.038 - 6.066: 99.4813% ( 1) 00:11:26.643 6.261 - 6.289: 99.4874% ( 1) 00:11:26.643 7.235 - 7.290: 99.4935% ( 1) 00:11:26.643 8.014 - 8.070: 99.4996% ( 1) 00:11:26.643 8.237 - 8.292: 99.5057% ( 1) 00:11:26.643 9.071 - 9.127: 99.5118% ( 1) 00:11:26.643 9.683 - 9.739: 99.5179% ( 1) 00:11:26.643 12.633 - 12.689: 99.5240% ( 1) 00:11:26.643 14.247 - 14.358: 99.5301% ( 1) 00:11:26.643 15.026 - 15.137: 99.5362% ( 1) 00:11:26.643 17.586 - 17.697: 99.5423% ( 1) 00:11:26.643 3034.602 - 3048.849: 99.5484% ( 1) 00:11:26.643 3989.148 - 4017.642: 99.9939% ( 73) 00:11:26.643 4160.111 - 4188.605: 100.0000% ( 1) 00:11:26.643 00:11:26.643 17:57:20 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@90 -- # aer_vfio_user /var/run/vfio-user/domain/vfio-user1/1 nqn.2019-07.io.spdk:cnode1 1 00:11:26.643 17:57:20 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@22 -- # local traddr=/var/run/vfio-user/domain/vfio-user1/1 00:11:26.643 17:57:20 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@23 -- # local subnqn=nqn.2019-07.io.spdk:cnode1 00:11:26.643 17:57:20 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@24 -- # local malloc_num=Malloc3 00:11:26.643 17:57:20 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:11:26.902 [ 00:11:26.902 { 00:11:26.902 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:11:26.902 "subtype": "Discovery", 00:11:26.902 "listen_addresses": [], 00:11:26.902 "allow_any_host": true, 00:11:26.902 "hosts": [] 00:11:26.902 }, 00:11:26.902 { 00:11:26.902 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:11:26.902 "subtype": "NVMe", 00:11:26.902 "listen_addresses": [ 00:11:26.902 { 00:11:26.902 "trtype": "VFIOUSER", 00:11:26.902 "adrfam": "IPv4", 00:11:26.902 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:11:26.902 "trsvcid": "0" 00:11:26.902 } 00:11:26.902 ], 00:11:26.902 "allow_any_host": true, 00:11:26.902 "hosts": [], 00:11:26.902 "serial_number": "SPDK1", 00:11:26.902 "model_number": "SPDK bdev Controller", 00:11:26.902 "max_namespaces": 32, 00:11:26.902 "min_cntlid": 1, 00:11:26.902 "max_cntlid": 65519, 00:11:26.902 "namespaces": [ 00:11:26.902 { 00:11:26.902 "nsid": 1, 00:11:26.902 "bdev_name": "Malloc1", 00:11:26.902 "name": "Malloc1", 00:11:26.902 "nguid": "D36E78CC2C2C4326A6A01BE2FFD8961C", 00:11:26.902 "uuid": "d36e78cc-2c2c-4326-a6a0-1be2ffd8961c" 00:11:26.902 } 00:11:26.902 ] 00:11:26.902 }, 00:11:26.902 { 00:11:26.902 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:11:26.902 "subtype": "NVMe", 00:11:26.902 "listen_addresses": [ 00:11:26.902 { 00:11:26.902 "trtype": "VFIOUSER", 00:11:26.902 "adrfam": "IPv4", 00:11:26.902 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:11:26.902 "trsvcid": "0" 00:11:26.902 } 00:11:26.902 ], 00:11:26.902 "allow_any_host": true, 00:11:26.902 "hosts": [], 00:11:26.902 "serial_number": "SPDK2", 00:11:26.902 "model_number": "SPDK bdev Controller", 00:11:26.902 "max_namespaces": 32, 00:11:26.902 "min_cntlid": 1, 00:11:26.902 "max_cntlid": 65519, 00:11:26.902 "namespaces": [ 00:11:26.902 { 00:11:26.902 "nsid": 1, 00:11:26.902 "bdev_name": "Malloc2", 00:11:26.902 "name": "Malloc2", 00:11:26.902 "nguid": "9BD4076E3A234916B3CE06FDCCE95459", 00:11:26.902 "uuid": "9bd4076e-3a23-4916-b3ce-06fdcce95459" 00:11:26.902 } 00:11:26.902 ] 00:11:26.902 } 00:11:26.902 ] 00:11:26.902 17:57:20 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@27 -- # AER_TOUCH_FILE=/tmp/aer_touch_file 00:11:26.902 17:57:20 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@34 -- # aerpid=513962 00:11:26.902 17:57:20 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@37 -- # waitforfile /tmp/aer_touch_file 00:11:26.902 17:57:20 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/aer/aer -r ' trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user1/1 subnqn:nqn.2019-07.io.spdk:cnode1' -n 2 -g -t /tmp/aer_touch_file 00:11:26.902 17:57:20 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@1265 -- # local i=0 00:11:26.902 17:57:20 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@1266 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:11:26.902 17:57:20 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@1272 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:11:26.902 17:57:20 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@1276 -- # return 0 00:11:26.902 17:57:20 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@38 -- # rm -f /tmp/aer_touch_file 00:11:26.902 17:57:20 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 --name Malloc3 00:11:26.902 EAL: No free 2048 kB hugepages reported on node 1 00:11:27.160 [2024-07-15 17:57:20.661629] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: enabling controller 00:11:27.161 Malloc3 00:11:27.161 17:57:20 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode1 Malloc3 -n 2 00:11:27.420 [2024-07-15 17:57:20.889369] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user1/1: disabling controller 00:11:27.420 17:57:20 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:11:27.420 Asynchronous Event Request test 00:11:27.420 Attaching to /var/run/vfio-user/domain/vfio-user1/1 00:11:27.420 Attached to /var/run/vfio-user/domain/vfio-user1/1 00:11:27.420 Registering asynchronous event callbacks... 00:11:27.420 Starting namespace attribute notice tests for all controllers... 00:11:27.420 /var/run/vfio-user/domain/vfio-user1/1: aer_cb for log page 4, aen_event_type: 0x02, aen_event_info: 0x00 00:11:27.420 aer_cb - Changed Namespace 00:11:27.420 Cleaning up... 00:11:27.420 [ 00:11:27.420 { 00:11:27.420 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:11:27.420 "subtype": "Discovery", 00:11:27.420 "listen_addresses": [], 00:11:27.420 "allow_any_host": true, 00:11:27.420 "hosts": [] 00:11:27.420 }, 00:11:27.420 { 00:11:27.420 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:11:27.420 "subtype": "NVMe", 00:11:27.420 "listen_addresses": [ 00:11:27.420 { 00:11:27.420 "trtype": "VFIOUSER", 00:11:27.420 "adrfam": "IPv4", 00:11:27.420 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:11:27.420 "trsvcid": "0" 00:11:27.420 } 00:11:27.420 ], 00:11:27.420 "allow_any_host": true, 00:11:27.420 "hosts": [], 00:11:27.420 "serial_number": "SPDK1", 00:11:27.420 "model_number": "SPDK bdev Controller", 00:11:27.420 "max_namespaces": 32, 00:11:27.420 "min_cntlid": 1, 00:11:27.420 "max_cntlid": 65519, 00:11:27.420 "namespaces": [ 00:11:27.420 { 00:11:27.420 "nsid": 1, 00:11:27.420 "bdev_name": "Malloc1", 00:11:27.420 "name": "Malloc1", 00:11:27.420 "nguid": "D36E78CC2C2C4326A6A01BE2FFD8961C", 00:11:27.420 "uuid": "d36e78cc-2c2c-4326-a6a0-1be2ffd8961c" 00:11:27.420 }, 00:11:27.420 { 00:11:27.420 "nsid": 2, 00:11:27.420 "bdev_name": "Malloc3", 00:11:27.420 "name": "Malloc3", 00:11:27.420 "nguid": "465C2DADFA0042F7AD71237783616747", 00:11:27.420 "uuid": "465c2dad-fa00-42f7-ad71-237783616747" 00:11:27.420 } 00:11:27.420 ] 00:11:27.420 }, 00:11:27.420 { 00:11:27.420 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:11:27.420 "subtype": "NVMe", 00:11:27.420 "listen_addresses": [ 00:11:27.420 { 00:11:27.420 "trtype": "VFIOUSER", 00:11:27.420 "adrfam": "IPv4", 00:11:27.420 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:11:27.420 "trsvcid": "0" 00:11:27.420 } 00:11:27.420 ], 00:11:27.420 "allow_any_host": true, 00:11:27.420 "hosts": [], 00:11:27.420 "serial_number": "SPDK2", 00:11:27.420 "model_number": "SPDK bdev Controller", 00:11:27.420 "max_namespaces": 32, 00:11:27.420 "min_cntlid": 1, 00:11:27.420 "max_cntlid": 65519, 00:11:27.420 "namespaces": [ 00:11:27.420 { 00:11:27.420 "nsid": 1, 00:11:27.420 "bdev_name": "Malloc2", 00:11:27.420 "name": "Malloc2", 00:11:27.420 "nguid": "9BD4076E3A234916B3CE06FDCCE95459", 00:11:27.420 "uuid": "9bd4076e-3a23-4916-b3ce-06fdcce95459" 00:11:27.420 } 00:11:27.420 ] 00:11:27.420 } 00:11:27.420 ] 00:11:27.420 17:57:21 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@44 -- # wait 513962 00:11:27.420 17:57:21 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@80 -- # for i in $(seq 1 $NUM_DEVICES) 00:11:27.420 17:57:21 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@81 -- # test_traddr=/var/run/vfio-user/domain/vfio-user2/2 00:11:27.420 17:57:21 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@82 -- # test_subnqn=nqn.2019-07.io.spdk:cnode2 00:11:27.420 17:57:21 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -g -L nvme -L nvme_vfio -L vfio_pci 00:11:27.420 [2024-07-15 17:57:21.116366] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:11:27.420 [2024-07-15 17:57:21.116392] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --single-file-segments --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid513981 ] 00:11:27.420 EAL: No free 2048 kB hugepages reported on node 1 00:11:27.420 [2024-07-15 17:57:21.142604] nvme_vfio_user.c: 259:nvme_vfio_ctrlr_scan: *DEBUG*: Scan controller : /var/run/vfio-user/domain/vfio-user2/2 00:11:27.681 [2024-07-15 17:57:21.152471] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 0, Size 0x2000, Offset 0x0, Flags 0xf, Cap offset 32 00:11:27.681 [2024-07-15 17:57:21.152493] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0x1000, Offset 0x1000, Map addr 0x7f56be1fb000 00:11:27.681 [2024-07-15 17:57:21.153468] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 1, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:11:27.681 [2024-07-15 17:57:21.154479] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 2, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:11:27.681 [2024-07-15 17:57:21.155488] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 3, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:11:27.681 [2024-07-15 17:57:21.156496] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 4, Size 0x2000, Offset 0x0, Flags 0x3, Cap offset 0 00:11:27.681 [2024-07-15 17:57:21.157500] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 5, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:11:27.681 [2024-07-15 17:57:21.158502] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 6, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:11:27.681 [2024-07-15 17:57:21.159508] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 7, Size 0x1000, Offset 0x0, Flags 0x3, Cap offset 0 00:11:27.681 [2024-07-15 17:57:21.160522] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 8, Size 0x0, Offset 0x0, Flags 0x0, Cap offset 0 00:11:27.681 [2024-07-15 17:57:21.161528] vfio_user_pci.c: 304:vfio_device_map_bars_and_config_region: *DEBUG*: Bar 9, Size 0xc000, Offset 0x0, Flags 0xf, Cap offset 32 00:11:27.681 [2024-07-15 17:57:21.161537] vfio_user_pci.c: 233:vfio_device_setup_sparse_mmaps: *DEBUG*: Sparse region 0, Size 0xb000, Offset 0x1000, Map addr 0x7f56be1f0000 00:11:27.681 [2024-07-15 17:57:21.162476] vfio_user_pci.c: 65:vfio_add_mr: *DEBUG*: Add memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:11:27.681 [2024-07-15 17:57:21.174997] vfio_user_pci.c: 386:spdk_vfio_user_setup: *DEBUG*: Device vfio-user0, Path /var/run/vfio-user/domain/vfio-user2/2/cntrl Setup Successfully 00:11:27.681 [2024-07-15 17:57:21.175018] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to connect adminq (no timeout) 00:11:27.681 [2024-07-15 17:57:21.177079] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x0, value 0x201e0100ff 00:11:27.681 [2024-07-15 17:57:21.177115] nvme_pcie_common.c: 132:nvme_pcie_qpair_construct: *INFO*: max_completions_cap = 64 num_trackers = 192 00:11:27.681 [2024-07-15 17:57:21.177180] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for connect adminq (no timeout) 00:11:27.681 [2024-07-15 17:57:21.177195] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to read vs (no timeout) 00:11:27.681 [2024-07-15 17:57:21.177200] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to read vs wait for vs (no timeout) 00:11:27.681 [2024-07-15 17:57:21.178077] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x8, value 0x10300 00:11:27.681 [2024-07-15 17:57:21.178087] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to read cap (no timeout) 00:11:27.681 [2024-07-15 17:57:21.178094] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to read cap wait for cap (no timeout) 00:11:27.681 [2024-07-15 17:57:21.182230] nvme_vfio_user.c: 103:nvme_vfio_ctrlr_get_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x0, value 0x201e0100ff 00:11:27.681 [2024-07-15 17:57:21.182239] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to check en (no timeout) 00:11:27.681 [2024-07-15 17:57:21.182246] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to check en wait for cc (timeout 15000 ms) 00:11:27.681 [2024-07-15 17:57:21.183112] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x0 00:11:27.681 [2024-07-15 17:57:21.183121] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:11:27.681 [2024-07-15 17:57:21.184119] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x1c, value 0x0 00:11:27.681 [2024-07-15 17:57:21.184127] nvme_ctrlr.c:3869:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] CC.EN = 0 && CSTS.RDY = 0 00:11:27.681 [2024-07-15 17:57:21.184131] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to controller is disabled (timeout 15000 ms) 00:11:27.681 [2024-07-15 17:57:21.184137] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:11:27.681 [2024-07-15 17:57:21.184242] nvme_ctrlr.c:4062:nvme_ctrlr_process_init: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] Setting CC.EN = 1 00:11:27.681 [2024-07-15 17:57:21.184247] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:11:27.681 [2024-07-15 17:57:21.184251] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x28, value 0x2000003c0000 00:11:27.681 [2024-07-15 17:57:21.185126] nvme_vfio_user.c: 61:nvme_vfio_ctrlr_set_reg_8: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x30, value 0x2000003be000 00:11:27.681 [2024-07-15 17:57:21.186131] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x24, value 0xff00ff 00:11:27.681 [2024-07-15 17:57:21.187146] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x460001 00:11:27.681 [2024-07-15 17:57:21.188155] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:11:27.681 [2024-07-15 17:57:21.188191] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:11:27.681 [2024-07-15 17:57:21.189165] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x1c, value 0x1 00:11:27.681 [2024-07-15 17:57:21.189173] nvme_ctrlr.c:3904:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:11:27.681 [2024-07-15 17:57:21.189177] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to reset admin queue (timeout 30000 ms) 00:11:27.681 [2024-07-15 17:57:21.189194] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to identify controller (no timeout) 00:11:27.681 [2024-07-15 17:57:21.189204] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for identify controller (timeout 30000 ms) 00:11:27.681 [2024-07-15 17:57:21.189215] nvme_pcie_common.c:1201:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:11:27.681 [2024-07-15 17:57:21.189219] nvme_pcie_common.c:1229:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:11:27.681 [2024-07-15 17:57:21.189234] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000001 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:11:27.681 [2024-07-15 17:57:21.193234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0001 p:1 m:0 dnr:0 00:11:27.681 [2024-07-15 17:57:21.193244] nvme_ctrlr.c:2053:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] transport max_xfer_size 131072 00:11:27.681 [2024-07-15 17:57:21.193251] nvme_ctrlr.c:2057:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] MDTS max_xfer_size 131072 00:11:27.681 [2024-07-15 17:57:21.193255] nvme_ctrlr.c:2060:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] CNTLID 0x0001 00:11:27.681 [2024-07-15 17:57:21.193259] nvme_ctrlr.c:2071:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] Identify CNTLID 0x0001 != Connect CNTLID 0x0000 00:11:27.681 [2024-07-15 17:57:21.193264] nvme_ctrlr.c:2084:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] transport max_sges 1 00:11:27.681 [2024-07-15 17:57:21.193269] nvme_ctrlr.c:2099:nvme_ctrlr_identify_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] fuses compare and write: 1 00:11:27.681 [2024-07-15 17:57:21.193273] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to configure AER (timeout 30000 ms) 00:11:27.681 [2024-07-15 17:57:21.193280] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for configure aer (timeout 30000 ms) 00:11:27.681 [2024-07-15 17:57:21.193289] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:191 cdw10:0000000b PRP1 0x0 PRP2 0x0 00:11:27.681 [2024-07-15 17:57:21.201231] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0002 p:1 m:0 dnr:0 00:11:27.681 [2024-07-15 17:57:21.201245] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:190 nsid:0 cdw10:00000000 cdw11:00000000 00:11:27.681 [2024-07-15 17:57:21.201252] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:189 nsid:0 cdw10:00000000 cdw11:00000000 00:11:27.681 [2024-07-15 17:57:21.201260] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:188 nsid:0 cdw10:00000000 cdw11:00000000 00:11:27.681 [2024-07-15 17:57:21.201267] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:187 nsid:0 cdw10:00000000 cdw11:00000000 00:11:27.682 [2024-07-15 17:57:21.201273] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to set keep alive timeout (timeout 30000 ms) 00:11:27.682 [2024-07-15 17:57:21.201280] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:11:27.682 [2024-07-15 17:57:21.201289] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:191 cdw10:0000000f PRP1 0x0 PRP2 0x0 00:11:27.682 [2024-07-15 17:57:21.209232] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0007 p:1 m:0 dnr:0 00:11:27.682 [2024-07-15 17:57:21.209240] nvme_ctrlr.c:3010:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] Controller adjusted keep alive timeout to 0 ms 00:11:27.682 [2024-07-15 17:57:21.209245] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to identify controller iocs specific (timeout 30000 ms) 00:11:27.682 [2024-07-15 17:57:21.209251] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to set number of queues (timeout 30000 ms) 00:11:27.682 [2024-07-15 17:57:21.209256] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for set number of queues (timeout 30000 ms) 00:11:27.682 [2024-07-15 17:57:21.209264] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:11:27.682 [2024-07-15 17:57:21.217229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:0008 p:1 m:0 dnr:0 00:11:27.682 [2024-07-15 17:57:21.217280] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to identify active ns (timeout 30000 ms) 00:11:27.682 [2024-07-15 17:57:21.217287] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for identify active ns (timeout 30000 ms) 00:11:27.682 [2024-07-15 17:57:21.217294] nvme_pcie_common.c:1201:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f9000 len:4096 00:11:27.682 [2024-07-15 17:57:21.217298] nvme_pcie_common.c:1229:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f9000 00:11:27.682 [2024-07-15 17:57:21.217304] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:0 cdw10:00000002 cdw11:00000000 PRP1 0x2000002f9000 PRP2 0x0 00:11:27.682 [2024-07-15 17:57:21.225231] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0009 p:1 m:0 dnr:0 00:11:27.682 [2024-07-15 17:57:21.225241] nvme_ctrlr.c:4693:spdk_nvme_ctrlr_get_ns: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] Namespace 1 was added 00:11:27.682 [2024-07-15 17:57:21.225250] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to identify ns (timeout 30000 ms) 00:11:27.682 [2024-07-15 17:57:21.225256] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for identify ns (timeout 30000 ms) 00:11:27.682 [2024-07-15 17:57:21.225263] nvme_pcie_common.c:1201:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:11:27.682 [2024-07-15 17:57:21.225267] nvme_pcie_common.c:1229:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:11:27.682 [2024-07-15 17:57:21.225272] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000000 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:11:27.682 [2024-07-15 17:57:21.233230] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000a p:1 m:0 dnr:0 00:11:27.682 [2024-07-15 17:57:21.233243] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to identify namespace id descriptors (timeout 30000 ms) 00:11:27.682 [2024-07-15 17:57:21.233250] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to wait for identify namespace id descriptors (timeout 30000 ms) 00:11:27.682 [2024-07-15 17:57:21.233259] nvme_pcie_common.c:1201:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:4096 00:11:27.682 [2024-07-15 17:57:21.233263] nvme_pcie_common.c:1229:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:11:27.682 [2024-07-15 17:57:21.233269] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:191 nsid:1 cdw10:00000003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:11:27.682 [2024-07-15 17:57:21.241229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000b p:1 m:0 dnr:0 00:11:27.682 [2024-07-15 17:57:21.241238] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to identify ns iocs specific (timeout 30000 ms) 00:11:27.682 [2024-07-15 17:57:21.241244] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to set supported log pages (timeout 30000 ms) 00:11:27.682 [2024-07-15 17:57:21.241252] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to set supported features (timeout 30000 ms) 00:11:27.682 [2024-07-15 17:57:21.241259] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to set host behavior support feature (timeout 30000 ms) 00:11:27.682 [2024-07-15 17:57:21.241263] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to set doorbell buffer config (timeout 30000 ms) 00:11:27.682 [2024-07-15 17:57:21.241268] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to set host ID (timeout 30000 ms) 00:11:27.682 [2024-07-15 17:57:21.241273] nvme_ctrlr.c:3110:nvme_ctrlr_set_host_id: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] NVMe-oF transport - not sending Set Features - Host ID 00:11:27.682 [2024-07-15 17:57:21.241277] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to transport ready (timeout 30000 ms) 00:11:27.682 [2024-07-15 17:57:21.241281] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] setting state to ready (no timeout) 00:11:27.682 [2024-07-15 17:57:21.241296] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ARBITRATION cid:191 cdw10:00000001 PRP1 0x0 PRP2 0x0 00:11:27.682 [2024-07-15 17:57:21.249231] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000c p:1 m:0 dnr:0 00:11:27.682 [2024-07-15 17:57:21.249246] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES POWER MANAGEMENT cid:191 cdw10:00000002 PRP1 0x0 PRP2 0x0 00:11:27.682 [2024-07-15 17:57:21.257229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000d p:1 m:0 dnr:0 00:11:27.682 [2024-07-15 17:57:21.257241] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES TEMPERATURE THRESHOLD cid:191 cdw10:00000004 PRP1 0x0 PRP2 0x0 00:11:27.682 [2024-07-15 17:57:21.265230] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:000e p:1 m:0 dnr:0 00:11:27.682 [2024-07-15 17:57:21.265242] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES NUMBER OF QUEUES cid:191 cdw10:00000007 PRP1 0x0 PRP2 0x0 00:11:27.682 [2024-07-15 17:57:21.273229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:7e007e sqhd:000f p:1 m:0 dnr:0 00:11:27.682 [2024-07-15 17:57:21.273245] nvme_pcie_common.c:1201:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f6000 len:8192 00:11:27.682 [2024-07-15 17:57:21.273249] nvme_pcie_common.c:1229:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f6000 00:11:27.682 [2024-07-15 17:57:21.273252] nvme_pcie_common.c:1238:nvme_pcie_prp_list_append: *DEBUG*: prp[0] = 0x2000002f7000 00:11:27.682 [2024-07-15 17:57:21.273255] nvme_pcie_common.c:1254:nvme_pcie_prp_list_append: *DEBUG*: prp2 = 0x2000002f7000 00:11:27.682 [2024-07-15 17:57:21.273261] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:191 nsid:ffffffff cdw10:07ff0001 cdw11:00000000 PRP1 0x2000002f6000 PRP2 0x2000002f7000 00:11:27.682 [2024-07-15 17:57:21.273270] nvme_pcie_common.c:1201:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fc000 len:512 00:11:27.682 [2024-07-15 17:57:21.273274] nvme_pcie_common.c:1229:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fc000 00:11:27.682 [2024-07-15 17:57:21.273280] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:186 nsid:ffffffff cdw10:007f0002 cdw11:00000000 PRP1 0x2000002fc000 PRP2 0x0 00:11:27.682 [2024-07-15 17:57:21.273286] nvme_pcie_common.c:1201:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002fb000 len:512 00:11:27.682 [2024-07-15 17:57:21.273290] nvme_pcie_common.c:1229:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002fb000 00:11:27.682 [2024-07-15 17:57:21.273296] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:185 nsid:ffffffff cdw10:007f0003 cdw11:00000000 PRP1 0x2000002fb000 PRP2 0x0 00:11:27.682 [2024-07-15 17:57:21.273303] nvme_pcie_common.c:1201:nvme_pcie_prp_list_append: *DEBUG*: prp_index:0 virt_addr:0x2000002f4000 len:4096 00:11:27.682 [2024-07-15 17:57:21.273306] nvme_pcie_common.c:1229:nvme_pcie_prp_list_append: *DEBUG*: prp1 = 0x2000002f4000 00:11:27.682 [2024-07-15 17:57:21.273312] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:184 nsid:ffffffff cdw10:03ff0005 cdw11:00000000 PRP1 0x2000002f4000 PRP2 0x0 00:11:27.682 [2024-07-15 17:57:21.281229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:191 cdw0:0 sqhd:0010 p:1 m:0 dnr:0 00:11:27.682 [2024-07-15 17:57:21.281242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:186 cdw0:0 sqhd:0011 p:1 m:0 dnr:0 00:11:27.682 [2024-07-15 17:57:21.281251] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:185 cdw0:0 sqhd:0012 p:1 m:0 dnr:0 00:11:27.682 [2024-07-15 17:57:21.281257] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0013 p:1 m:0 dnr:0 00:11:27.682 ===================================================== 00:11:27.682 NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:11:27.682 ===================================================== 00:11:27.682 Controller Capabilities/Features 00:11:27.682 ================================ 00:11:27.682 Vendor ID: 4e58 00:11:27.682 Subsystem Vendor ID: 4e58 00:11:27.682 Serial Number: SPDK2 00:11:27.682 Model Number: SPDK bdev Controller 00:11:27.682 Firmware Version: 24.09 00:11:27.682 Recommended Arb Burst: 6 00:11:27.682 IEEE OUI Identifier: 8d 6b 50 00:11:27.682 Multi-path I/O 00:11:27.682 May have multiple subsystem ports: Yes 00:11:27.682 May have multiple controllers: Yes 00:11:27.682 Associated with SR-IOV VF: No 00:11:27.682 Max Data Transfer Size: 131072 00:11:27.682 Max Number of Namespaces: 32 00:11:27.682 Max Number of I/O Queues: 127 00:11:27.682 NVMe Specification Version (VS): 1.3 00:11:27.682 NVMe Specification Version (Identify): 1.3 00:11:27.682 Maximum Queue Entries: 256 00:11:27.682 Contiguous Queues Required: Yes 00:11:27.682 Arbitration Mechanisms Supported 00:11:27.682 Weighted Round Robin: Not Supported 00:11:27.682 Vendor Specific: Not Supported 00:11:27.682 Reset Timeout: 15000 ms 00:11:27.682 Doorbell Stride: 4 bytes 00:11:27.682 NVM Subsystem Reset: Not Supported 00:11:27.682 Command Sets Supported 00:11:27.682 NVM Command Set: Supported 00:11:27.682 Boot Partition: Not Supported 00:11:27.682 Memory Page Size Minimum: 4096 bytes 00:11:27.682 Memory Page Size Maximum: 4096 bytes 00:11:27.682 Persistent Memory Region: Not Supported 00:11:27.682 Optional Asynchronous Events Supported 00:11:27.682 Namespace Attribute Notices: Supported 00:11:27.682 Firmware Activation Notices: Not Supported 00:11:27.682 ANA Change Notices: Not Supported 00:11:27.682 PLE Aggregate Log Change Notices: Not Supported 00:11:27.682 LBA Status Info Alert Notices: Not Supported 00:11:27.682 EGE Aggregate Log Change Notices: Not Supported 00:11:27.682 Normal NVM Subsystem Shutdown event: Not Supported 00:11:27.682 Zone Descriptor Change Notices: Not Supported 00:11:27.683 Discovery Log Change Notices: Not Supported 00:11:27.683 Controller Attributes 00:11:27.683 128-bit Host Identifier: Supported 00:11:27.683 Non-Operational Permissive Mode: Not Supported 00:11:27.683 NVM Sets: Not Supported 00:11:27.683 Read Recovery Levels: Not Supported 00:11:27.683 Endurance Groups: Not Supported 00:11:27.683 Predictable Latency Mode: Not Supported 00:11:27.683 Traffic Based Keep ALive: Not Supported 00:11:27.683 Namespace Granularity: Not Supported 00:11:27.683 SQ Associations: Not Supported 00:11:27.683 UUID List: Not Supported 00:11:27.683 Multi-Domain Subsystem: Not Supported 00:11:27.683 Fixed Capacity Management: Not Supported 00:11:27.683 Variable Capacity Management: Not Supported 00:11:27.683 Delete Endurance Group: Not Supported 00:11:27.683 Delete NVM Set: Not Supported 00:11:27.683 Extended LBA Formats Supported: Not Supported 00:11:27.683 Flexible Data Placement Supported: Not Supported 00:11:27.683 00:11:27.683 Controller Memory Buffer Support 00:11:27.683 ================================ 00:11:27.683 Supported: No 00:11:27.683 00:11:27.683 Persistent Memory Region Support 00:11:27.683 ================================ 00:11:27.683 Supported: No 00:11:27.683 00:11:27.683 Admin Command Set Attributes 00:11:27.683 ============================ 00:11:27.683 Security Send/Receive: Not Supported 00:11:27.683 Format NVM: Not Supported 00:11:27.683 Firmware Activate/Download: Not Supported 00:11:27.683 Namespace Management: Not Supported 00:11:27.683 Device Self-Test: Not Supported 00:11:27.683 Directives: Not Supported 00:11:27.683 NVMe-MI: Not Supported 00:11:27.683 Virtualization Management: Not Supported 00:11:27.683 Doorbell Buffer Config: Not Supported 00:11:27.683 Get LBA Status Capability: Not Supported 00:11:27.683 Command & Feature Lockdown Capability: Not Supported 00:11:27.683 Abort Command Limit: 4 00:11:27.683 Async Event Request Limit: 4 00:11:27.683 Number of Firmware Slots: N/A 00:11:27.683 Firmware Slot 1 Read-Only: N/A 00:11:27.683 Firmware Activation Without Reset: N/A 00:11:27.683 Multiple Update Detection Support: N/A 00:11:27.683 Firmware Update Granularity: No Information Provided 00:11:27.683 Per-Namespace SMART Log: No 00:11:27.683 Asymmetric Namespace Access Log Page: Not Supported 00:11:27.683 Subsystem NQN: nqn.2019-07.io.spdk:cnode2 00:11:27.683 Command Effects Log Page: Supported 00:11:27.683 Get Log Page Extended Data: Supported 00:11:27.683 Telemetry Log Pages: Not Supported 00:11:27.683 Persistent Event Log Pages: Not Supported 00:11:27.683 Supported Log Pages Log Page: May Support 00:11:27.683 Commands Supported & Effects Log Page: Not Supported 00:11:27.683 Feature Identifiers & Effects Log Page:May Support 00:11:27.683 NVMe-MI Commands & Effects Log Page: May Support 00:11:27.683 Data Area 4 for Telemetry Log: Not Supported 00:11:27.683 Error Log Page Entries Supported: 128 00:11:27.683 Keep Alive: Supported 00:11:27.683 Keep Alive Granularity: 10000 ms 00:11:27.683 00:11:27.683 NVM Command Set Attributes 00:11:27.683 ========================== 00:11:27.683 Submission Queue Entry Size 00:11:27.683 Max: 64 00:11:27.683 Min: 64 00:11:27.683 Completion Queue Entry Size 00:11:27.683 Max: 16 00:11:27.683 Min: 16 00:11:27.683 Number of Namespaces: 32 00:11:27.683 Compare Command: Supported 00:11:27.683 Write Uncorrectable Command: Not Supported 00:11:27.683 Dataset Management Command: Supported 00:11:27.683 Write Zeroes Command: Supported 00:11:27.683 Set Features Save Field: Not Supported 00:11:27.683 Reservations: Not Supported 00:11:27.683 Timestamp: Not Supported 00:11:27.683 Copy: Supported 00:11:27.683 Volatile Write Cache: Present 00:11:27.683 Atomic Write Unit (Normal): 1 00:11:27.683 Atomic Write Unit (PFail): 1 00:11:27.683 Atomic Compare & Write Unit: 1 00:11:27.683 Fused Compare & Write: Supported 00:11:27.683 Scatter-Gather List 00:11:27.683 SGL Command Set: Supported (Dword aligned) 00:11:27.683 SGL Keyed: Not Supported 00:11:27.683 SGL Bit Bucket Descriptor: Not Supported 00:11:27.683 SGL Metadata Pointer: Not Supported 00:11:27.683 Oversized SGL: Not Supported 00:11:27.683 SGL Metadata Address: Not Supported 00:11:27.683 SGL Offset: Not Supported 00:11:27.683 Transport SGL Data Block: Not Supported 00:11:27.683 Replay Protected Memory Block: Not Supported 00:11:27.683 00:11:27.683 Firmware Slot Information 00:11:27.683 ========================= 00:11:27.683 Active slot: 1 00:11:27.683 Slot 1 Firmware Revision: 24.09 00:11:27.683 00:11:27.683 00:11:27.683 Commands Supported and Effects 00:11:27.683 ============================== 00:11:27.683 Admin Commands 00:11:27.683 -------------- 00:11:27.683 Get Log Page (02h): Supported 00:11:27.683 Identify (06h): Supported 00:11:27.683 Abort (08h): Supported 00:11:27.683 Set Features (09h): Supported 00:11:27.683 Get Features (0Ah): Supported 00:11:27.683 Asynchronous Event Request (0Ch): Supported 00:11:27.683 Keep Alive (18h): Supported 00:11:27.683 I/O Commands 00:11:27.683 ------------ 00:11:27.683 Flush (00h): Supported LBA-Change 00:11:27.683 Write (01h): Supported LBA-Change 00:11:27.683 Read (02h): Supported 00:11:27.683 Compare (05h): Supported 00:11:27.683 Write Zeroes (08h): Supported LBA-Change 00:11:27.683 Dataset Management (09h): Supported LBA-Change 00:11:27.683 Copy (19h): Supported LBA-Change 00:11:27.683 00:11:27.683 Error Log 00:11:27.683 ========= 00:11:27.683 00:11:27.683 Arbitration 00:11:27.683 =========== 00:11:27.683 Arbitration Burst: 1 00:11:27.683 00:11:27.683 Power Management 00:11:27.683 ================ 00:11:27.683 Number of Power States: 1 00:11:27.683 Current Power State: Power State #0 00:11:27.683 Power State #0: 00:11:27.683 Max Power: 0.00 W 00:11:27.683 Non-Operational State: Operational 00:11:27.683 Entry Latency: Not Reported 00:11:27.683 Exit Latency: Not Reported 00:11:27.683 Relative Read Throughput: 0 00:11:27.683 Relative Read Latency: 0 00:11:27.683 Relative Write Throughput: 0 00:11:27.683 Relative Write Latency: 0 00:11:27.683 Idle Power: Not Reported 00:11:27.683 Active Power: Not Reported 00:11:27.683 Non-Operational Permissive Mode: Not Supported 00:11:27.683 00:11:27.683 Health Information 00:11:27.683 ================== 00:11:27.683 Critical Warnings: 00:11:27.683 Available Spare Space: OK 00:11:27.683 Temperature: OK 00:11:27.683 Device Reliability: OK 00:11:27.683 Read Only: No 00:11:27.683 Volatile Memory Backup: OK 00:11:27.683 Current Temperature: 0 Kelvin (-273 Celsius) 00:11:27.683 Temperature Threshold: 0 Kelvin (-273 Celsius) 00:11:27.683 Available Spare: 0% 00:11:27.683 Available Sp[2024-07-15 17:57:21.281343] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ERROR_RECOVERY cid:184 cdw10:00000005 PRP1 0x0 PRP2 0x0 00:11:27.683 [2024-07-15 17:57:21.289232] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: SUCCESS (00/00) qid:0 cid:184 cdw0:0 sqhd:0014 p:1 m:0 dnr:0 00:11:27.683 [2024-07-15 17:57:21.289263] nvme_ctrlr.c:4357:nvme_ctrlr_destruct_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] Prepare to destruct SSD 00:11:27.683 [2024-07-15 17:57:21.289271] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:190 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:27.683 [2024-07-15 17:57:21.289277] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:189 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:27.683 [2024-07-15 17:57:21.289283] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:188 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:27.683 [2024-07-15 17:57:21.289288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:187 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:11:27.683 [2024-07-15 17:57:21.289326] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x460001 00:11:27.683 [2024-07-15 17:57:21.289336] nvme_vfio_user.c: 49:nvme_vfio_ctrlr_set_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x14, value 0x464001 00:11:27.684 [2024-07-15 17:57:21.290329] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:11:27.684 [2024-07-15 17:57:21.290372] nvme_ctrlr.c:1147:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] RTD3E = 0 us 00:11:27.684 [2024-07-15 17:57:21.290378] nvme_ctrlr.c:1150:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] shutdown timeout = 10000 ms 00:11:27.684 [2024-07-15 17:57:21.291337] nvme_vfio_user.c: 83:nvme_vfio_ctrlr_get_reg_4: *DEBUG*: ctrlr /var/run/vfio-user/domain/vfio-user2/2: offset 0x1c, value 0x9 00:11:27.684 [2024-07-15 17:57:21.291348] nvme_ctrlr.c:1269:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [/var/run/vfio-user/domain/vfio-user2/2] shutdown complete in 0 milliseconds 00:11:27.684 [2024-07-15 17:57:21.291393] vfio_user_pci.c: 399:spdk_vfio_user_release: *DEBUG*: Release file /var/run/vfio-user/domain/vfio-user2/2/cntrl 00:11:27.684 [2024-07-15 17:57:21.294230] vfio_user_pci.c: 96:vfio_remove_mr: *DEBUG*: Remove memory region: FD 10, VADDR 0x200000200000, IOVA 0x200000200000, Size 0x200000 00:11:27.684 are Threshold: 0% 00:11:27.684 Life Percentage Used: 0% 00:11:27.684 Data Units Read: 0 00:11:27.684 Data Units Written: 0 00:11:27.684 Host Read Commands: 0 00:11:27.684 Host Write Commands: 0 00:11:27.684 Controller Busy Time: 0 minutes 00:11:27.684 Power Cycles: 0 00:11:27.684 Power On Hours: 0 hours 00:11:27.684 Unsafe Shutdowns: 0 00:11:27.684 Unrecoverable Media Errors: 0 00:11:27.684 Lifetime Error Log Entries: 0 00:11:27.684 Warning Temperature Time: 0 minutes 00:11:27.684 Critical Temperature Time: 0 minutes 00:11:27.684 00:11:27.684 Number of Queues 00:11:27.684 ================ 00:11:27.684 Number of I/O Submission Queues: 127 00:11:27.684 Number of I/O Completion Queues: 127 00:11:27.684 00:11:27.684 Active Namespaces 00:11:27.684 ================= 00:11:27.684 Namespace ID:1 00:11:27.684 Error Recovery Timeout: Unlimited 00:11:27.684 Command Set Identifier: NVM (00h) 00:11:27.684 Deallocate: Supported 00:11:27.684 Deallocated/Unwritten Error: Not Supported 00:11:27.684 Deallocated Read Value: Unknown 00:11:27.684 Deallocate in Write Zeroes: Not Supported 00:11:27.684 Deallocated Guard Field: 0xFFFF 00:11:27.684 Flush: Supported 00:11:27.684 Reservation: Supported 00:11:27.684 Namespace Sharing Capabilities: Multiple Controllers 00:11:27.684 Size (in LBAs): 131072 (0GiB) 00:11:27.684 Capacity (in LBAs): 131072 (0GiB) 00:11:27.684 Utilization (in LBAs): 131072 (0GiB) 00:11:27.684 NGUID: 9BD4076E3A234916B3CE06FDCCE95459 00:11:27.684 UUID: 9bd4076e-3a23-4916-b3ce-06fdcce95459 00:11:27.684 Thin Provisioning: Not Supported 00:11:27.684 Per-NS Atomic Units: Yes 00:11:27.684 Atomic Boundary Size (Normal): 0 00:11:27.684 Atomic Boundary Size (PFail): 0 00:11:27.684 Atomic Boundary Offset: 0 00:11:27.684 Maximum Single Source Range Length: 65535 00:11:27.684 Maximum Copy Length: 65535 00:11:27.684 Maximum Source Range Count: 1 00:11:27.684 NGUID/EUI64 Never Reused: No 00:11:27.684 Namespace Write Protected: No 00:11:27.684 Number of LBA Formats: 1 00:11:27.684 Current LBA Format: LBA Format #00 00:11:27.684 LBA Format #00: Data Size: 512 Metadata Size: 0 00:11:27.684 00:11:27.684 17:57:21 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -s 256 -g -q 128 -o 4096 -w read -t 5 -c 0x2 00:11:27.684 EAL: No free 2048 kB hugepages reported on node 1 00:11:27.942 [2024-07-15 17:57:21.507187] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:11:33.251 Initializing NVMe Controllers 00:11:33.251 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:11:33.251 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 with lcore 1 00:11:33.251 Initialization complete. Launching workers. 00:11:33.251 ======================================================== 00:11:33.251 Latency(us) 00:11:33.251 Device Information : IOPS MiB/s Average min max 00:11:33.251 VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 from core 1: 39937.73 156.01 3204.58 946.84 6660.40 00:11:33.251 ======================================================== 00:11:33.251 Total : 39937.73 156.01 3204.58 946.84 6660.40 00:11:33.251 00:11:33.251 [2024-07-15 17:57:26.614460] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:11:33.251 17:57:26 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@85 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -s 256 -g -q 128 -o 4096 -w write -t 5 -c 0x2 00:11:33.251 EAL: No free 2048 kB hugepages reported on node 1 00:11:33.251 [2024-07-15 17:57:26.829054] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:11:38.524 Initializing NVMe Controllers 00:11:38.524 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:11:38.524 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 with lcore 1 00:11:38.524 Initialization complete. Launching workers. 00:11:38.524 ======================================================== 00:11:38.524 Latency(us) 00:11:38.524 Device Information : IOPS MiB/s Average min max 00:11:38.524 VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) NSID 1 from core 1: 39935.04 156.00 3204.80 962.46 7472.81 00:11:38.524 ======================================================== 00:11:38.524 Total : 39935.04 156.00 3204.80 962.46 7472.81 00:11:38.524 00:11:38.524 [2024-07-15 17:57:31.847140] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:11:38.524 17:57:31 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -g -q 32 -o 4096 -w randrw -M 50 -t 5 -c 0xE 00:11:38.524 EAL: No free 2048 kB hugepages reported on node 1 00:11:38.524 [2024-07-15 17:57:32.043557] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:11:43.795 [2024-07-15 17:57:37.176314] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:11:43.795 Initializing NVMe Controllers 00:11:43.795 Attaching to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:11:43.795 Attached to NVMe over Fabrics controller at /var/run/vfio-user/domain/vfio-user2/2:: nqn.2019-07.io.spdk:cnode2 00:11:43.795 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) with lcore 1 00:11:43.795 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) with lcore 2 00:11:43.795 Associating VFIOUSER (/var/run/vfio-user/domain/vfio-user2/2) with lcore 3 00:11:43.795 Initialization complete. Launching workers. 00:11:43.795 Starting thread on core 2 00:11:43.795 Starting thread on core 3 00:11:43.795 Starting thread on core 1 00:11:43.795 17:57:37 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -t 3 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -d 256 -g 00:11:43.795 EAL: No free 2048 kB hugepages reported on node 1 00:11:43.795 [2024-07-15 17:57:37.459660] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:11:47.082 [2024-07-15 17:57:40.508430] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:11:47.082 Initializing NVMe Controllers 00:11:47.082 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:11:47.082 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:11:47.082 Associating SPDK bdev Controller (SPDK2 ) with lcore 0 00:11:47.082 Associating SPDK bdev Controller (SPDK2 ) with lcore 1 00:11:47.082 Associating SPDK bdev Controller (SPDK2 ) with lcore 2 00:11:47.082 Associating SPDK bdev Controller (SPDK2 ) with lcore 3 00:11:47.082 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration run with configuration: 00:11:47.082 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/arbitration -q 64 -s 131072 -w randrw -M 50 -l 0 -t 3 -c 0xf -m 0 -a 0 -b 0 -n 100000 -i -1 00:11:47.082 Initialization complete. Launching workers. 00:11:47.082 Starting thread on core 1 with urgent priority queue 00:11:47.082 Starting thread on core 2 with urgent priority queue 00:11:47.082 Starting thread on core 3 with urgent priority queue 00:11:47.082 Starting thread on core 0 with urgent priority queue 00:11:47.082 SPDK bdev Controller (SPDK2 ) core 0: 7947.00 IO/s 12.58 secs/100000 ios 00:11:47.082 SPDK bdev Controller (SPDK2 ) core 1: 8038.00 IO/s 12.44 secs/100000 ios 00:11:47.082 SPDK bdev Controller (SPDK2 ) core 2: 9108.67 IO/s 10.98 secs/100000 ios 00:11:47.082 SPDK bdev Controller (SPDK2 ) core 3: 10190.00 IO/s 9.81 secs/100000 ios 00:11:47.082 ======================================================== 00:11:47.082 00:11:47.082 17:57:40 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/hello_world -d 256 -g -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' 00:11:47.082 EAL: No free 2048 kB hugepages reported on node 1 00:11:47.082 [2024-07-15 17:57:40.780605] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:11:47.082 Initializing NVMe Controllers 00:11:47.082 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:11:47.082 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:11:47.082 Namespace ID: 1 size: 0GB 00:11:47.082 Initialization complete. 00:11:47.082 INFO: using host memory buffer for IO 00:11:47.082 Hello world! 00:11:47.082 [2024-07-15 17:57:40.790665] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:11:47.350 17:57:40 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/overhead/overhead -o 4096 -t 1 -H -g -d 256 -r 'trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' 00:11:47.350 EAL: No free 2048 kB hugepages reported on node 1 00:11:47.350 [2024-07-15 17:57:41.062794] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:11:48.731 Initializing NVMe Controllers 00:11:48.731 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:11:48.731 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:11:48.731 Initialization complete. Launching workers. 00:11:48.731 submit (in ns) avg, min, max = 8317.2, 3263.5, 3999486.1 00:11:48.731 complete (in ns) avg, min, max = 18947.9, 1773.0, 3998367.8 00:11:48.731 00:11:48.731 Submit histogram 00:11:48.731 ================ 00:11:48.731 Range in us Cumulative Count 00:11:48.731 3.256 - 3.270: 0.0120% ( 2) 00:11:48.731 3.283 - 3.297: 0.0722% ( 10) 00:11:48.731 3.297 - 3.311: 0.2226% ( 25) 00:11:48.731 3.311 - 3.325: 0.5173% ( 49) 00:11:48.731 3.325 - 3.339: 1.0467% ( 88) 00:11:48.731 3.339 - 3.353: 2.0753% ( 171) 00:11:48.731 3.353 - 3.367: 4.6379% ( 426) 00:11:48.731 3.367 - 3.381: 9.5344% ( 814) 00:11:48.731 3.381 - 3.395: 15.4235% ( 979) 00:11:48.731 3.395 - 3.409: 21.5772% ( 1023) 00:11:48.731 3.409 - 3.423: 28.0017% ( 1068) 00:11:48.731 3.423 - 3.437: 33.4396% ( 904) 00:11:48.731 3.437 - 3.450: 38.5707% ( 853) 00:11:48.731 3.450 - 3.464: 43.9124% ( 888) 00:11:48.731 3.464 - 3.478: 48.9533% ( 838) 00:11:48.731 3.478 - 3.492: 53.0318% ( 678) 00:11:48.731 3.492 - 3.506: 57.2305% ( 698) 00:11:48.731 3.506 - 3.520: 63.3000% ( 1009) 00:11:48.731 3.520 - 3.534: 68.9966% ( 947) 00:11:48.731 3.534 - 3.548: 73.1292% ( 687) 00:11:48.731 3.548 - 3.562: 77.8633% ( 787) 00:11:48.731 3.562 - 3.590: 84.6547% ( 1129) 00:11:48.731 3.590 - 3.617: 87.0067% ( 391) 00:11:48.731 3.617 - 3.645: 87.9211% ( 152) 00:11:48.731 3.645 - 3.673: 89.0279% ( 184) 00:11:48.731 3.673 - 3.701: 90.7664% ( 289) 00:11:48.731 3.701 - 3.729: 92.4687% ( 283) 00:11:48.731 3.729 - 3.757: 93.9786% ( 251) 00:11:48.731 3.757 - 3.784: 95.6629% ( 280) 00:11:48.731 3.784 - 3.812: 97.2871% ( 270) 00:11:48.731 3.812 - 3.840: 98.2856% ( 166) 00:11:48.731 3.840 - 3.868: 98.9112% ( 104) 00:11:48.731 3.868 - 3.896: 99.3203% ( 68) 00:11:48.731 3.896 - 3.923: 99.4406% ( 20) 00:11:48.731 3.923 - 3.951: 99.4887% ( 8) 00:11:48.731 3.951 - 3.979: 99.5308% ( 7) 00:11:48.731 3.979 - 4.007: 99.5428% ( 2) 00:11:48.731 4.007 - 4.035: 99.5549% ( 2) 00:11:48.731 4.090 - 4.118: 99.5609% ( 1) 00:11:48.731 4.118 - 4.146: 99.5669% ( 1) 00:11:48.731 4.146 - 4.174: 99.5729% ( 1) 00:11:48.731 4.591 - 4.619: 99.5789% ( 1) 00:11:48.731 5.009 - 5.037: 99.5970% ( 3) 00:11:48.731 5.064 - 5.092: 99.6090% ( 2) 00:11:48.731 5.092 - 5.120: 99.6150% ( 1) 00:11:48.731 5.120 - 5.148: 99.6270% ( 2) 00:11:48.731 5.148 - 5.176: 99.6331% ( 1) 00:11:48.731 5.176 - 5.203: 99.6391% ( 1) 00:11:48.731 5.287 - 5.315: 99.6451% ( 1) 00:11:48.731 5.315 - 5.343: 99.6571% ( 2) 00:11:48.731 5.343 - 5.370: 99.6631% ( 1) 00:11:48.731 5.370 - 5.398: 99.6752% ( 2) 00:11:48.731 5.398 - 5.426: 99.6812% ( 1) 00:11:48.731 5.426 - 5.454: 99.6872% ( 1) 00:11:48.731 5.454 - 5.482: 99.7052% ( 3) 00:11:48.731 5.537 - 5.565: 99.7113% ( 1) 00:11:48.731 5.565 - 5.593: 99.7173% ( 1) 00:11:48.731 5.621 - 5.649: 99.7233% ( 1) 00:11:48.731 5.649 - 5.677: 99.7293% ( 1) 00:11:48.731 5.677 - 5.704: 99.7413% ( 2) 00:11:48.731 5.760 - 5.788: 99.7534% ( 2) 00:11:48.731 5.899 - 5.927: 99.7594% ( 1) 00:11:48.731 5.927 - 5.955: 99.7654% ( 1) 00:11:48.731 6.010 - 6.038: 99.7714% ( 1) 00:11:48.731 6.317 - 6.344: 99.7774% ( 1) 00:11:48.731 6.344 - 6.372: 99.7834% ( 1) 00:11:48.731 6.372 - 6.400: 99.7895% ( 1) 00:11:48.731 6.595 - 6.623: 99.8015% ( 2) 00:11:48.731 6.623 - 6.650: 99.8075% ( 1) 00:11:48.731 6.734 - 6.762: 99.8135% ( 1) 00:11:48.731 6.790 - 6.817: 99.8195% ( 1) 00:11:48.731 6.929 - 6.957: 99.8256% ( 1) 00:11:48.731 7.096 - 7.123: 99.8316% ( 1) 00:11:48.731 7.123 - 7.179: 99.8436% ( 2) 00:11:48.731 7.179 - 7.235: 99.8496% ( 1) 00:11:48.731 7.513 - 7.569: 99.8556% ( 1) 00:11:48.731 8.181 - 8.237: 99.8616% ( 1) 00:11:48.731 8.960 - 9.016: 99.8677% ( 1) 00:11:48.731 19.033 - 19.144: 99.8737% ( 1) 00:11:48.731 [2024-07-15 17:57:42.155283] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:11:48.731 30.497 - 30.720: 99.8797% ( 1) 00:11:48.731 3989.148 - 4017.642: 100.0000% ( 20) 00:11:48.731 00:11:48.731 Complete histogram 00:11:48.731 ================== 00:11:48.731 Range in us Cumulative Count 00:11:48.731 1.767 - 1.774: 0.0060% ( 1) 00:11:48.731 1.774 - 1.781: 0.0120% ( 1) 00:11:48.731 1.781 - 1.795: 0.0301% ( 3) 00:11:48.731 1.795 - 1.809: 0.0541% ( 4) 00:11:48.731 1.809 - 1.823: 3.5912% ( 588) 00:11:48.731 1.823 - 1.837: 16.7409% ( 2186) 00:11:48.731 1.837 - 1.850: 20.2358% ( 581) 00:11:48.731 1.850 - 1.864: 21.6374% ( 233) 00:11:48.731 1.864 - 1.878: 39.8881% ( 3034) 00:11:48.731 1.878 - 1.892: 84.5344% ( 7422) 00:11:48.731 1.892 - 1.906: 93.6838% ( 1521) 00:11:48.731 1.906 - 1.920: 95.9456% ( 376) 00:11:48.731 1.920 - 1.934: 96.6675% ( 120) 00:11:48.731 1.934 - 1.948: 97.2690% ( 100) 00:11:48.731 1.948 - 1.962: 98.0931% ( 137) 00:11:48.731 1.962 - 1.976: 98.7789% ( 114) 00:11:48.731 1.976 - 1.990: 99.0676% ( 48) 00:11:48.731 1.990 - 2.003: 99.1218% ( 9) 00:11:48.731 2.003 - 2.017: 99.1639% ( 7) 00:11:48.731 2.017 - 2.031: 99.2000% ( 6) 00:11:48.731 2.031 - 2.045: 99.2300% ( 5) 00:11:48.731 2.045 - 2.059: 99.2661% ( 6) 00:11:48.731 2.059 - 2.073: 99.2842% ( 3) 00:11:48.731 2.073 - 2.087: 99.2902% ( 1) 00:11:48.731 2.101 - 2.115: 99.3082% ( 3) 00:11:48.731 2.115 - 2.129: 99.3323% ( 4) 00:11:48.731 2.129 - 2.143: 99.3443% ( 2) 00:11:48.731 2.143 - 2.157: 99.3564% ( 2) 00:11:48.732 2.157 - 2.170: 99.3624% ( 1) 00:11:48.732 2.170 - 2.184: 99.3804% ( 3) 00:11:48.732 2.268 - 2.282: 99.3864% ( 1) 00:11:48.732 2.296 - 2.310: 99.3924% ( 1) 00:11:48.732 2.518 - 2.532: 99.3985% ( 1) 00:11:48.732 2.560 - 2.574: 99.4045% ( 1) 00:11:48.732 3.325 - 3.339: 99.4105% ( 1) 00:11:48.732 3.339 - 3.353: 99.4165% ( 1) 00:11:48.732 3.617 - 3.645: 99.4225% ( 1) 00:11:48.732 3.701 - 3.729: 99.4285% ( 1) 00:11:48.732 3.729 - 3.757: 99.4346% ( 1) 00:11:48.732 3.923 - 3.951: 99.4406% ( 1) 00:11:48.732 3.951 - 3.979: 99.4466% ( 1) 00:11:48.732 4.063 - 4.090: 99.4526% ( 1) 00:11:48.732 4.090 - 4.118: 99.4646% ( 2) 00:11:48.732 4.146 - 4.174: 99.4706% ( 1) 00:11:48.732 4.202 - 4.230: 99.4767% ( 1) 00:11:48.732 4.230 - 4.257: 99.4887% ( 2) 00:11:48.732 4.341 - 4.369: 99.4947% ( 1) 00:11:48.732 4.424 - 4.452: 99.5007% ( 1) 00:11:48.732 4.452 - 4.480: 99.5067% ( 1) 00:11:48.732 4.480 - 4.508: 99.5128% ( 1) 00:11:48.732 4.675 - 4.703: 99.5188% ( 1) 00:11:48.732 4.730 - 4.758: 99.5248% ( 1) 00:11:48.732 4.953 - 4.981: 99.5308% ( 1) 00:11:48.732 5.398 - 5.426: 99.5368% ( 1) 00:11:48.732 6.177 - 6.205: 99.5428% ( 1) 00:11:48.732 7.847 - 7.903: 99.5488% ( 1) 00:11:48.732 8.626 - 8.682: 99.5549% ( 1) 00:11:48.732 12.355 - 12.410: 99.5609% ( 1) 00:11:48.732 25.711 - 25.823: 99.5669% ( 1) 00:11:48.732 154.045 - 154.936: 99.5729% ( 1) 00:11:48.732 3989.148 - 4017.642: 100.0000% ( 71) 00:11:48.732 00:11:48.732 17:57:42 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@90 -- # aer_vfio_user /var/run/vfio-user/domain/vfio-user2/2 nqn.2019-07.io.spdk:cnode2 2 00:11:48.732 17:57:42 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@22 -- # local traddr=/var/run/vfio-user/domain/vfio-user2/2 00:11:48.732 17:57:42 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@23 -- # local subnqn=nqn.2019-07.io.spdk:cnode2 00:11:48.732 17:57:42 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@24 -- # local malloc_num=Malloc4 00:11:48.732 17:57:42 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:11:48.732 [ 00:11:48.732 { 00:11:48.732 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:11:48.732 "subtype": "Discovery", 00:11:48.732 "listen_addresses": [], 00:11:48.732 "allow_any_host": true, 00:11:48.732 "hosts": [] 00:11:48.732 }, 00:11:48.732 { 00:11:48.732 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:11:48.732 "subtype": "NVMe", 00:11:48.732 "listen_addresses": [ 00:11:48.732 { 00:11:48.732 "trtype": "VFIOUSER", 00:11:48.732 "adrfam": "IPv4", 00:11:48.732 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:11:48.732 "trsvcid": "0" 00:11:48.732 } 00:11:48.732 ], 00:11:48.732 "allow_any_host": true, 00:11:48.732 "hosts": [], 00:11:48.732 "serial_number": "SPDK1", 00:11:48.732 "model_number": "SPDK bdev Controller", 00:11:48.732 "max_namespaces": 32, 00:11:48.732 "min_cntlid": 1, 00:11:48.732 "max_cntlid": 65519, 00:11:48.732 "namespaces": [ 00:11:48.732 { 00:11:48.732 "nsid": 1, 00:11:48.732 "bdev_name": "Malloc1", 00:11:48.732 "name": "Malloc1", 00:11:48.732 "nguid": "D36E78CC2C2C4326A6A01BE2FFD8961C", 00:11:48.732 "uuid": "d36e78cc-2c2c-4326-a6a0-1be2ffd8961c" 00:11:48.732 }, 00:11:48.732 { 00:11:48.732 "nsid": 2, 00:11:48.732 "bdev_name": "Malloc3", 00:11:48.732 "name": "Malloc3", 00:11:48.732 "nguid": "465C2DADFA0042F7AD71237783616747", 00:11:48.732 "uuid": "465c2dad-fa00-42f7-ad71-237783616747" 00:11:48.732 } 00:11:48.732 ] 00:11:48.732 }, 00:11:48.732 { 00:11:48.732 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:11:48.732 "subtype": "NVMe", 00:11:48.732 "listen_addresses": [ 00:11:48.732 { 00:11:48.732 "trtype": "VFIOUSER", 00:11:48.732 "adrfam": "IPv4", 00:11:48.732 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:11:48.732 "trsvcid": "0" 00:11:48.732 } 00:11:48.732 ], 00:11:48.732 "allow_any_host": true, 00:11:48.732 "hosts": [], 00:11:48.732 "serial_number": "SPDK2", 00:11:48.732 "model_number": "SPDK bdev Controller", 00:11:48.732 "max_namespaces": 32, 00:11:48.732 "min_cntlid": 1, 00:11:48.732 "max_cntlid": 65519, 00:11:48.732 "namespaces": [ 00:11:48.732 { 00:11:48.732 "nsid": 1, 00:11:48.732 "bdev_name": "Malloc2", 00:11:48.732 "name": "Malloc2", 00:11:48.732 "nguid": "9BD4076E3A234916B3CE06FDCCE95459", 00:11:48.732 "uuid": "9bd4076e-3a23-4916-b3ce-06fdcce95459" 00:11:48.732 } 00:11:48.732 ] 00:11:48.732 } 00:11:48.732 ] 00:11:48.732 17:57:42 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@27 -- # AER_TOUCH_FILE=/tmp/aer_touch_file 00:11:48.732 17:57:42 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@34 -- # aerpid=517486 00:11:48.732 17:57:42 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@37 -- # waitforfile /tmp/aer_touch_file 00:11:48.732 17:57:42 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/aer/aer -r ' trtype:VFIOUSER traddr:/var/run/vfio-user/domain/vfio-user2/2 subnqn:nqn.2019-07.io.spdk:cnode2' -n 2 -g -t /tmp/aer_touch_file 00:11:48.732 17:57:42 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@1265 -- # local i=0 00:11:48.732 17:57:42 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@1266 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:11:48.732 17:57:42 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@1272 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:11:48.732 17:57:42 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@1276 -- # return 0 00:11:48.732 17:57:42 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@38 -- # rm -f /tmp/aer_touch_file 00:11:48.732 17:57:42 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 --name Malloc4 00:11:48.732 EAL: No free 2048 kB hugepages reported on node 1 00:11:48.990 [2024-07-15 17:57:42.530624] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: enabling controller 00:11:48.990 Malloc4 00:11:48.990 17:57:42 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode2 Malloc4 -n 2 00:11:49.249 [2024-07-15 17:57:42.771421] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user/domain/vfio-user2/2: disabling controller 00:11:49.250 17:57:42 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_get_subsystems 00:11:49.250 Asynchronous Event Request test 00:11:49.250 Attaching to /var/run/vfio-user/domain/vfio-user2/2 00:11:49.250 Attached to /var/run/vfio-user/domain/vfio-user2/2 00:11:49.250 Registering asynchronous event callbacks... 00:11:49.250 Starting namespace attribute notice tests for all controllers... 00:11:49.250 /var/run/vfio-user/domain/vfio-user2/2: aer_cb for log page 4, aen_event_type: 0x02, aen_event_info: 0x00 00:11:49.250 aer_cb - Changed Namespace 00:11:49.250 Cleaning up... 00:11:49.250 [ 00:11:49.250 { 00:11:49.250 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:11:49.250 "subtype": "Discovery", 00:11:49.250 "listen_addresses": [], 00:11:49.250 "allow_any_host": true, 00:11:49.250 "hosts": [] 00:11:49.250 }, 00:11:49.250 { 00:11:49.250 "nqn": "nqn.2019-07.io.spdk:cnode1", 00:11:49.250 "subtype": "NVMe", 00:11:49.250 "listen_addresses": [ 00:11:49.250 { 00:11:49.250 "trtype": "VFIOUSER", 00:11:49.250 "adrfam": "IPv4", 00:11:49.250 "traddr": "/var/run/vfio-user/domain/vfio-user1/1", 00:11:49.250 "trsvcid": "0" 00:11:49.250 } 00:11:49.250 ], 00:11:49.250 "allow_any_host": true, 00:11:49.250 "hosts": [], 00:11:49.250 "serial_number": "SPDK1", 00:11:49.250 "model_number": "SPDK bdev Controller", 00:11:49.250 "max_namespaces": 32, 00:11:49.250 "min_cntlid": 1, 00:11:49.250 "max_cntlid": 65519, 00:11:49.250 "namespaces": [ 00:11:49.250 { 00:11:49.250 "nsid": 1, 00:11:49.250 "bdev_name": "Malloc1", 00:11:49.250 "name": "Malloc1", 00:11:49.250 "nguid": "D36E78CC2C2C4326A6A01BE2FFD8961C", 00:11:49.250 "uuid": "d36e78cc-2c2c-4326-a6a0-1be2ffd8961c" 00:11:49.250 }, 00:11:49.250 { 00:11:49.250 "nsid": 2, 00:11:49.250 "bdev_name": "Malloc3", 00:11:49.250 "name": "Malloc3", 00:11:49.250 "nguid": "465C2DADFA0042F7AD71237783616747", 00:11:49.250 "uuid": "465c2dad-fa00-42f7-ad71-237783616747" 00:11:49.250 } 00:11:49.250 ] 00:11:49.250 }, 00:11:49.250 { 00:11:49.250 "nqn": "nqn.2019-07.io.spdk:cnode2", 00:11:49.250 "subtype": "NVMe", 00:11:49.250 "listen_addresses": [ 00:11:49.250 { 00:11:49.250 "trtype": "VFIOUSER", 00:11:49.250 "adrfam": "IPv4", 00:11:49.250 "traddr": "/var/run/vfio-user/domain/vfio-user2/2", 00:11:49.250 "trsvcid": "0" 00:11:49.250 } 00:11:49.250 ], 00:11:49.250 "allow_any_host": true, 00:11:49.250 "hosts": [], 00:11:49.250 "serial_number": "SPDK2", 00:11:49.250 "model_number": "SPDK bdev Controller", 00:11:49.250 "max_namespaces": 32, 00:11:49.250 "min_cntlid": 1, 00:11:49.250 "max_cntlid": 65519, 00:11:49.250 "namespaces": [ 00:11:49.250 { 00:11:49.250 "nsid": 1, 00:11:49.250 "bdev_name": "Malloc2", 00:11:49.250 "name": "Malloc2", 00:11:49.250 "nguid": "9BD4076E3A234916B3CE06FDCCE95459", 00:11:49.250 "uuid": "9bd4076e-3a23-4916-b3ce-06fdcce95459" 00:11:49.250 }, 00:11:49.250 { 00:11:49.250 "nsid": 2, 00:11:49.250 "bdev_name": "Malloc4", 00:11:49.250 "name": "Malloc4", 00:11:49.250 "nguid": "BA37E41416A344B68B2A5EC5C9912AB3", 00:11:49.250 "uuid": "ba37e414-16a3-44b6-8b2a-5ec5c9912ab3" 00:11:49.250 } 00:11:49.250 ] 00:11:49.250 } 00:11:49.250 ] 00:11:49.509 17:57:42 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@44 -- # wait 517486 00:11:49.509 17:57:42 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@105 -- # stop_nvmf_vfio_user 00:11:49.509 17:57:42 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@95 -- # killprocess 509286 00:11:49.509 17:57:42 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@948 -- # '[' -z 509286 ']' 00:11:49.509 17:57:42 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@952 -- # kill -0 509286 00:11:49.509 17:57:42 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@953 -- # uname 00:11:49.509 17:57:42 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:11:49.509 17:57:42 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 509286 00:11:49.509 17:57:43 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:11:49.509 17:57:43 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:11:49.509 17:57:43 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@966 -- # echo 'killing process with pid 509286' 00:11:49.509 killing process with pid 509286 00:11:49.509 17:57:43 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@967 -- # kill 509286 00:11:49.509 17:57:43 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@972 -- # wait 509286 00:11:49.768 17:57:43 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@97 -- # rm -rf /var/run/vfio-user 00:11:49.768 17:57:43 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:11:49.768 17:57:43 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@108 -- # setup_nvmf_vfio_user --interrupt-mode '-M -I' 00:11:49.768 17:57:43 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@51 -- # local nvmf_app_args=--interrupt-mode 00:11:49.768 17:57:43 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@52 -- # local 'transport_args=-M -I' 00:11:49.768 17:57:43 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@55 -- # nvmfpid=517670 00:11:49.768 17:57:43 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@57 -- # echo 'Process pid: 517670' 00:11:49.768 Process pid: 517670 00:11:49.768 17:57:43 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m '[0,1,2,3]' --interrupt-mode 00:11:49.768 17:57:43 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@59 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:11:49.768 17:57:43 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@60 -- # waitforlisten 517670 00:11:49.768 17:57:43 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@829 -- # '[' -z 517670 ']' 00:11:49.768 17:57:43 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:49.768 17:57:43 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@834 -- # local max_retries=100 00:11:49.768 17:57:43 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:49.768 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:49.768 17:57:43 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@838 -- # xtrace_disable 00:11:49.768 17:57:43 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@10 -- # set +x 00:11:49.768 [2024-07-15 17:57:43.336614] thread.c:2948:spdk_interrupt_mode_enable: *NOTICE*: Set SPDK running in interrupt mode. 00:11:49.768 [2024-07-15 17:57:43.337472] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:11:49.768 [2024-07-15 17:57:43.337510] [ DPDK EAL parameters: nvmf -l 0,1,2,3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:49.768 EAL: No free 2048 kB hugepages reported on node 1 00:11:49.768 [2024-07-15 17:57:43.390680] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:11:49.768 [2024-07-15 17:57:43.458743] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:49.768 [2024-07-15 17:57:43.458783] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:49.768 [2024-07-15 17:57:43.458790] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:49.768 [2024-07-15 17:57:43.458795] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:49.768 [2024-07-15 17:57:43.458800] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:49.768 [2024-07-15 17:57:43.458892] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:11:49.768 [2024-07-15 17:57:43.459008] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:11:49.768 [2024-07-15 17:57:43.459073] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:11:49.768 [2024-07-15 17:57:43.459074] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:11:50.026 [2024-07-15 17:57:43.535158] thread.c:2099:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_000) to intr mode from intr mode. 00:11:50.027 [2024-07-15 17:57:43.535317] thread.c:2099:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_001) to intr mode from intr mode. 00:11:50.027 [2024-07-15 17:57:43.535421] thread.c:2099:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_002) to intr mode from intr mode. 00:11:50.027 [2024-07-15 17:57:43.535642] thread.c:2099:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (app_thread) to intr mode from intr mode. 00:11:50.027 [2024-07-15 17:57:43.535847] thread.c:2099:spdk_thread_set_interrupt_mode: *NOTICE*: Set spdk_thread (nvmf_tgt_poll_group_003) to intr mode from intr mode. 00:11:50.594 17:57:44 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:11:50.594 17:57:44 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@862 -- # return 0 00:11:50.594 17:57:44 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@62 -- # sleep 1 00:11:51.533 17:57:45 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t VFIOUSER -M -I 00:11:51.792 17:57:45 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@66 -- # mkdir -p /var/run/vfio-user 00:11:51.792 17:57:45 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # seq 1 2 00:11:51.792 17:57:45 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:11:51.792 17:57:45 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user1/1 00:11:51.792 17:57:45 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:11:51.792 Malloc1 00:11:52.051 17:57:45 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode1 -a -s SPDK1 00:11:52.051 17:57:45 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode1 Malloc1 00:11:52.310 17:57:45 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode1 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user1/1 -s 0 00:11:52.310 17:57:46 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@68 -- # for i in $(seq 1 $NUM_DEVICES) 00:11:52.310 17:57:46 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@69 -- # mkdir -p /var/run/vfio-user/domain/vfio-user2/2 00:11:52.310 17:57:46 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc2 00:11:52.569 Malloc2 00:11:52.569 17:57:46 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2019-07.io.spdk:cnode2 -a -s SPDK2 00:11:52.827 17:57:46 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2019-07.io.spdk:cnode2 Malloc2 00:11:53.086 17:57:46 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@74 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2019-07.io.spdk:cnode2 -t VFIOUSER -a /var/run/vfio-user/domain/vfio-user2/2 -s 0 00:11:53.086 17:57:46 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@109 -- # stop_nvmf_vfio_user 00:11:53.086 17:57:46 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@95 -- # killprocess 517670 00:11:53.086 17:57:46 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@948 -- # '[' -z 517670 ']' 00:11:53.086 17:57:46 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@952 -- # kill -0 517670 00:11:53.087 17:57:46 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@953 -- # uname 00:11:53.087 17:57:46 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:11:53.087 17:57:46 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 517670 00:11:53.087 17:57:46 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:11:53.087 17:57:46 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:11:53.087 17:57:46 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@966 -- # echo 'killing process with pid 517670' 00:11:53.087 killing process with pid 517670 00:11:53.087 17:57:46 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@967 -- # kill 517670 00:11:53.087 17:57:46 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@972 -- # wait 517670 00:11:53.346 17:57:47 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@97 -- # rm -rf /var/run/vfio-user 00:11:53.346 17:57:47 nvmf_tcp.nvmf_vfio_user -- target/nvmf_vfio_user.sh@99 -- # trap - SIGINT SIGTERM EXIT 00:11:53.346 00:11:53.346 real 0m51.361s 00:11:53.346 user 3m16.922s 00:11:53.346 sys 0m3.096s 00:11:53.346 17:57:47 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@1124 -- # xtrace_disable 00:11:53.346 17:57:47 nvmf_tcp.nvmf_vfio_user -- common/autotest_common.sh@10 -- # set +x 00:11:53.346 ************************************ 00:11:53.346 END TEST nvmf_vfio_user 00:11:53.346 ************************************ 00:11:53.346 17:57:47 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:11:53.346 17:57:47 nvmf_tcp -- nvmf/nvmf.sh@42 -- # run_test nvmf_vfio_user_nvme_compliance /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance/compliance.sh --transport=tcp 00:11:53.346 17:57:47 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:11:53.346 17:57:47 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:11:53.346 17:57:47 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:11:53.606 ************************************ 00:11:53.606 START TEST nvmf_vfio_user_nvme_compliance 00:11:53.606 ************************************ 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance/compliance.sh --transport=tcp 00:11:53.606 * Looking for test storage... 00:11:53.606 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@7 -- # uname -s 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:11:53.606 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- paths/export.sh@5 -- # export PATH 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@47 -- # : 0 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- nvmf/common.sh@51 -- # have_pci_nics=0 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@11 -- # MALLOC_BDEV_SIZE=64 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@14 -- # export TEST_TRANSPORT=VFIOUSER 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@14 -- # TEST_TRANSPORT=VFIOUSER 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@16 -- # rm -rf /var/run/vfio-user 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@20 -- # nvmfpid=518426 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@21 -- # echo 'Process pid: 518426' 00:11:53.607 Process pid: 518426 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@23 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x7 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@24 -- # waitforlisten 518426 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@829 -- # '[' -z 518426 ']' 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@834 -- # local max_retries=100 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:53.607 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@838 -- # xtrace_disable 00:11:53.607 17:57:47 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:11:53.607 [2024-07-15 17:57:47.248868] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:11:53.607 [2024-07-15 17:57:47.248917] [ DPDK EAL parameters: nvmf -c 0x7 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:11:53.607 EAL: No free 2048 kB hugepages reported on node 1 00:11:53.607 [2024-07-15 17:57:47.303947] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:11:53.866 [2024-07-15 17:57:47.383611] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:11:53.866 [2024-07-15 17:57:47.383643] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:11:53.866 [2024-07-15 17:57:47.383651] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:11:53.866 [2024-07-15 17:57:47.383656] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:11:53.866 [2024-07-15 17:57:47.383662] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:11:53.866 [2024-07-15 17:57:47.383703] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:11:53.866 [2024-07-15 17:57:47.383800] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:11:53.866 [2024-07-15 17:57:47.383801] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:11:54.435 17:57:48 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:11:54.435 17:57:48 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@862 -- # return 0 00:11:54.435 17:57:48 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@26 -- # sleep 1 00:11:55.373 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@28 -- # nqn=nqn.2021-09.io.spdk:cnode0 00:11:55.373 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@29 -- # traddr=/var/run/vfio-user 00:11:55.373 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@31 -- # rpc_cmd nvmf_create_transport -t VFIOUSER 00:11:55.373 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@559 -- # xtrace_disable 00:11:55.373 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:11:55.373 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:11:55.373 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@33 -- # mkdir -p /var/run/vfio-user 00:11:55.373 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@35 -- # rpc_cmd bdev_malloc_create 64 512 -b malloc0 00:11:55.373 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@559 -- # xtrace_disable 00:11:55.373 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:11:55.632 malloc0 00:11:55.632 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:11:55.632 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@36 -- # rpc_cmd nvmf_create_subsystem nqn.2021-09.io.spdk:cnode0 -a -s spdk -m 32 00:11:55.632 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@559 -- # xtrace_disable 00:11:55.632 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:11:55.632 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:11:55.632 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@37 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2021-09.io.spdk:cnode0 malloc0 00:11:55.632 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@559 -- # xtrace_disable 00:11:55.632 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:11:55.632 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:11:55.632 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@38 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2021-09.io.spdk:cnode0 -t VFIOUSER -a /var/run/vfio-user -s 0 00:11:55.632 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@559 -- # xtrace_disable 00:11:55.632 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:11:55.632 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:11:55.632 17:57:49 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/compliance/nvme_compliance -g -r 'trtype:VFIOUSER traddr:/var/run/vfio-user subnqn:nqn.2021-09.io.spdk:cnode0' 00:11:55.632 EAL: No free 2048 kB hugepages reported on node 1 00:11:55.632 00:11:55.632 00:11:55.632 CUnit - A unit testing framework for C - Version 2.1-3 00:11:55.632 http://cunit.sourceforge.net/ 00:11:55.632 00:11:55.632 00:11:55.632 Suite: nvme_compliance 00:11:55.632 Test: admin_identify_ctrlr_verify_dptr ...[2024-07-15 17:57:49.288633] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:11:55.632 [2024-07-15 17:57:49.289950] vfio_user.c: 804:nvme_cmd_map_prps: *ERROR*: no PRP2, 3072 remaining 00:11:55.632 [2024-07-15 17:57:49.289964] vfio_user.c:5511:map_admin_cmd_req: *ERROR*: /var/run/vfio-user: map Admin Opc 6 failed 00:11:55.632 [2024-07-15 17:57:49.289970] vfio_user.c:5604:handle_cmd_req: *ERROR*: /var/run/vfio-user: process NVMe command opc 0x6 failed 00:11:55.632 [2024-07-15 17:57:49.292654] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:11:55.632 passed 00:11:55.891 Test: admin_identify_ctrlr_verify_fused ...[2024-07-15 17:57:49.372209] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:11:55.891 [2024-07-15 17:57:49.375231] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:11:55.891 passed 00:11:55.891 Test: admin_identify_ns ...[2024-07-15 17:57:49.455223] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:11:55.891 [2024-07-15 17:57:49.517240] ctrlr.c:2729:_nvmf_ctrlr_get_ns_safe: *ERROR*: Identify Namespace for invalid NSID 0 00:11:55.891 [2024-07-15 17:57:49.525233] ctrlr.c:2729:_nvmf_ctrlr_get_ns_safe: *ERROR*: Identify Namespace for invalid NSID 4294967295 00:11:55.891 [2024-07-15 17:57:49.546335] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:11:55.891 passed 00:11:56.151 Test: admin_get_features_mandatory_features ...[2024-07-15 17:57:49.623493] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:11:56.151 [2024-07-15 17:57:49.626520] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:11:56.151 passed 00:11:56.151 Test: admin_get_features_optional_features ...[2024-07-15 17:57:49.705029] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:11:56.151 [2024-07-15 17:57:49.708050] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:11:56.151 passed 00:11:56.151 Test: admin_set_features_number_of_queues ...[2024-07-15 17:57:49.785657] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:11:56.410 [2024-07-15 17:57:49.894321] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:11:56.410 passed 00:11:56.410 Test: admin_get_log_page_mandatory_logs ...[2024-07-15 17:57:49.968513] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:11:56.410 [2024-07-15 17:57:49.972538] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:11:56.410 passed 00:11:56.410 Test: admin_get_log_page_with_lpo ...[2024-07-15 17:57:50.049548] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:11:56.410 [2024-07-15 17:57:50.121242] ctrlr.c:2677:nvmf_ctrlr_get_log_page: *ERROR*: Get log page: offset (516) > len (512) 00:11:56.410 [2024-07-15 17:57:50.134292] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:11:56.670 passed 00:11:56.670 Test: fabric_property_get ...[2024-07-15 17:57:50.208526] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:11:56.670 [2024-07-15 17:57:50.212464] vfio_user.c:5604:handle_cmd_req: *ERROR*: /var/run/vfio-user: process NVMe command opc 0x7f failed 00:11:56.670 [2024-07-15 17:57:50.213562] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:11:56.670 passed 00:11:56.670 Test: admin_delete_io_sq_use_admin_qid ...[2024-07-15 17:57:50.292058] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:11:56.670 [2024-07-15 17:57:50.293300] vfio_user.c:2309:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:0 does not exist 00:11:56.670 [2024-07-15 17:57:50.295090] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:11:56.670 passed 00:11:56.670 Test: admin_delete_io_sq_delete_sq_twice ...[2024-07-15 17:57:50.374759] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:11:56.928 [2024-07-15 17:57:50.459241] vfio_user.c:2309:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:1 does not exist 00:11:56.928 [2024-07-15 17:57:50.475236] vfio_user.c:2309:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:1 does not exist 00:11:56.928 [2024-07-15 17:57:50.480386] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:11:56.928 passed 00:11:56.928 Test: admin_delete_io_cq_use_admin_qid ...[2024-07-15 17:57:50.554519] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:11:56.928 [2024-07-15 17:57:50.555752] vfio_user.c:2309:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O cqid:0 does not exist 00:11:56.928 [2024-07-15 17:57:50.557542] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:11:56.928 passed 00:11:56.928 Test: admin_delete_io_cq_delete_cq_first ...[2024-07-15 17:57:50.637812] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:11:57.187 [2024-07-15 17:57:50.711236] vfio_user.c:2319:handle_del_io_q: *ERROR*: /var/run/vfio-user: the associated SQ must be deleted first 00:11:57.187 [2024-07-15 17:57:50.735235] vfio_user.c:2309:handle_del_io_q: *ERROR*: /var/run/vfio-user: I/O sqid:1 does not exist 00:11:57.187 [2024-07-15 17:57:50.740340] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:11:57.187 passed 00:11:57.187 Test: admin_create_io_cq_verify_iv_pc ...[2024-07-15 17:57:50.819317] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:11:57.187 [2024-07-15 17:57:50.820551] vfio_user.c:2158:handle_create_io_cq: *ERROR*: /var/run/vfio-user: IV is too big 00:11:57.187 [2024-07-15 17:57:50.820574] vfio_user.c:2152:handle_create_io_cq: *ERROR*: /var/run/vfio-user: non-PC CQ not supported 00:11:57.187 [2024-07-15 17:57:50.822339] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:11:57.187 passed 00:11:57.187 Test: admin_create_io_sq_verify_qsize_cqid ...[2024-07-15 17:57:50.899325] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:11:57.446 [2024-07-15 17:57:50.993248] vfio_user.c:2240:handle_create_io_q: *ERROR*: /var/run/vfio-user: invalid I/O queue size 1 00:11:57.446 [2024-07-15 17:57:51.001232] vfio_user.c:2240:handle_create_io_q: *ERROR*: /var/run/vfio-user: invalid I/O queue size 257 00:11:57.446 [2024-07-15 17:57:51.009229] vfio_user.c:2038:handle_create_io_sq: *ERROR*: /var/run/vfio-user: invalid cqid:0 00:11:57.446 [2024-07-15 17:57:51.017230] vfio_user.c:2038:handle_create_io_sq: *ERROR*: /var/run/vfio-user: invalid cqid:128 00:11:57.446 [2024-07-15 17:57:51.046315] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:11:57.446 passed 00:11:57.446 Test: admin_create_io_sq_verify_pc ...[2024-07-15 17:57:51.121458] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:11:57.446 [2024-07-15 17:57:51.138237] vfio_user.c:2051:handle_create_io_sq: *ERROR*: /var/run/vfio-user: non-PC SQ not supported 00:11:57.446 [2024-07-15 17:57:51.155623] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:11:57.705 passed 00:11:57.705 Test: admin_create_io_qp_max_qps ...[2024-07-15 17:57:51.235145] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:11:58.641 [2024-07-15 17:57:52.346234] nvme_ctrlr.c:5465:spdk_nvme_ctrlr_alloc_qid: *ERROR*: [/var/run/vfio-user] No free I/O queue IDs 00:11:59.210 [2024-07-15 17:57:52.721352] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:11:59.210 passed 00:11:59.210 Test: admin_create_io_sq_shared_cq ...[2024-07-15 17:57:52.800481] vfio_user.c:2836:enable_ctrlr: *NOTICE*: /var/run/vfio-user: enabling controller 00:11:59.210 [2024-07-15 17:57:52.933236] vfio_user.c:2319:handle_del_io_q: *ERROR*: /var/run/vfio-user: the associated SQ must be deleted first 00:11:59.470 [2024-07-15 17:57:52.970295] vfio_user.c:2798:disable_ctrlr: *NOTICE*: /var/run/vfio-user: disabling controller 00:11:59.470 passed 00:11:59.470 00:11:59.470 Run Summary: Type Total Ran Passed Failed Inactive 00:11:59.470 suites 1 1 n/a 0 0 00:11:59.470 tests 18 18 18 0 0 00:11:59.470 asserts 360 360 360 0 n/a 00:11:59.470 00:11:59.470 Elapsed time = 1.512 seconds 00:11:59.470 17:57:53 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@42 -- # killprocess 518426 00:11:59.470 17:57:53 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@948 -- # '[' -z 518426 ']' 00:11:59.470 17:57:53 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@952 -- # kill -0 518426 00:11:59.470 17:57:53 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@953 -- # uname 00:11:59.470 17:57:53 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:11:59.470 17:57:53 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 518426 00:11:59.470 17:57:53 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:11:59.470 17:57:53 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:11:59.470 17:57:53 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@966 -- # echo 'killing process with pid 518426' 00:11:59.470 killing process with pid 518426 00:11:59.470 17:57:53 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@967 -- # kill 518426 00:11:59.470 17:57:53 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@972 -- # wait 518426 00:11:59.729 17:57:53 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@44 -- # rm -rf /var/run/vfio-user 00:11:59.729 17:57:53 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- compliance/compliance.sh@46 -- # trap - SIGINT SIGTERM EXIT 00:11:59.730 00:11:59.730 real 0m6.180s 00:11:59.730 user 0m17.654s 00:11:59.730 sys 0m0.449s 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@1124 -- # xtrace_disable 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_nvme_compliance -- common/autotest_common.sh@10 -- # set +x 00:11:59.730 ************************************ 00:11:59.730 END TEST nvmf_vfio_user_nvme_compliance 00:11:59.730 ************************************ 00:11:59.730 17:57:53 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:11:59.730 17:57:53 nvmf_tcp -- nvmf/nvmf.sh@43 -- # run_test nvmf_vfio_user_fuzz /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/vfio_user_fuzz.sh --transport=tcp 00:11:59.730 17:57:53 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:11:59.730 17:57:53 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:11:59.730 17:57:53 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:11:59.730 ************************************ 00:11:59.730 START TEST nvmf_vfio_user_fuzz 00:11:59.730 ************************************ 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/vfio_user_fuzz.sh --transport=tcp 00:11:59.730 * Looking for test storage... 00:11:59.730 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@7 -- # uname -s 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- paths/export.sh@5 -- # export PATH 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@47 -- # : 0 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- nvmf/common.sh@51 -- # have_pci_nics=0 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@12 -- # MALLOC_BDEV_SIZE=64 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@15 -- # nqn=nqn.2021-09.io.spdk:cnode0 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@16 -- # traddr=/var/run/vfio-user 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@18 -- # export TEST_TRANSPORT=VFIOUSER 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@18 -- # TEST_TRANSPORT=VFIOUSER 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@20 -- # rm -rf /var/run/vfio-user 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@24 -- # nvmfpid=519464 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@25 -- # echo 'Process pid: 519464' 00:11:59.730 Process pid: 519464 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@27 -- # trap 'killprocess $nvmfpid; exit 1' SIGINT SIGTERM EXIT 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@28 -- # waitforlisten 519464 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@829 -- # '[' -z 519464 ']' 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@834 -- # local max_retries=100 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:11:59.730 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@838 -- # xtrace_disable 00:11:59.730 17:57:53 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:12:00.666 17:57:54 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:12:00.666 17:57:54 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@862 -- # return 0 00:12:00.666 17:57:54 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@30 -- # sleep 1 00:12:01.606 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@32 -- # rpc_cmd nvmf_create_transport -t VFIOUSER 00:12:01.606 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@559 -- # xtrace_disable 00:12:01.606 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:12:01.606 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:12:01.606 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@34 -- # mkdir -p /var/run/vfio-user 00:12:01.606 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@36 -- # rpc_cmd bdev_malloc_create 64 512 -b malloc0 00:12:01.606 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@559 -- # xtrace_disable 00:12:01.606 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:12:01.865 malloc0 00:12:01.865 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:12:01.865 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@37 -- # rpc_cmd nvmf_create_subsystem nqn.2021-09.io.spdk:cnode0 -a -s spdk 00:12:01.865 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@559 -- # xtrace_disable 00:12:01.865 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:12:01.865 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:12:01.865 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@38 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2021-09.io.spdk:cnode0 malloc0 00:12:01.865 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@559 -- # xtrace_disable 00:12:01.865 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:12:01.865 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:12:01.865 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@39 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2021-09.io.spdk:cnode0 -t VFIOUSER -a /var/run/vfio-user -s 0 00:12:01.865 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@559 -- # xtrace_disable 00:12:01.865 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:12:01.866 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:12:01.866 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@41 -- # trid='trtype:VFIOUSER subnqn:nqn.2021-09.io.spdk:cnode0 traddr:/var/run/vfio-user' 00:12:01.866 17:57:55 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/fuzz/nvme_fuzz/nvme_fuzz -m 0x2 -t 30 -S 123456 -F 'trtype:VFIOUSER subnqn:nqn.2021-09.io.spdk:cnode0 traddr:/var/run/vfio-user' -N -a 00:12:34.051 Fuzzing completed. Shutting down the fuzz application 00:12:34.051 00:12:34.051 Dumping successful admin opcodes: 00:12:34.051 8, 9, 10, 24, 00:12:34.051 Dumping successful io opcodes: 00:12:34.051 0, 00:12:34.051 NS: 0x200003a1ef00 I/O qp, Total commands completed: 1140176, total successful commands: 4493, random_seed: 2689611328 00:12:34.051 NS: 0x200003a1ef00 admin qp, Total commands completed: 282403, total successful commands: 2274, random_seed: 2297833024 00:12:34.051 17:58:25 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@44 -- # rpc_cmd nvmf_delete_subsystem nqn.2021-09.io.spdk:cnode0 00:12:34.051 17:58:25 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@559 -- # xtrace_disable 00:12:34.051 17:58:25 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:12:34.051 17:58:25 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:12:34.051 17:58:25 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@46 -- # killprocess 519464 00:12:34.051 17:58:25 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@948 -- # '[' -z 519464 ']' 00:12:34.051 17:58:25 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@952 -- # kill -0 519464 00:12:34.051 17:58:25 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@953 -- # uname 00:12:34.051 17:58:25 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:12:34.051 17:58:25 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 519464 00:12:34.051 17:58:25 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:12:34.051 17:58:25 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:12:34.051 17:58:25 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@966 -- # echo 'killing process with pid 519464' 00:12:34.051 killing process with pid 519464 00:12:34.051 17:58:25 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@967 -- # kill 519464 00:12:34.051 17:58:25 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@972 -- # wait 519464 00:12:34.051 17:58:26 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@48 -- # rm -rf /var/run/vfio-user /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/vfio_user_fuzz_log.txt /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/vfio_user_fuzz_tgt_output.txt 00:12:34.051 17:58:26 nvmf_tcp.nvmf_vfio_user_fuzz -- target/vfio_user_fuzz.sh@50 -- # trap - SIGINT SIGTERM EXIT 00:12:34.051 00:12:34.051 real 0m32.786s 00:12:34.051 user 0m35.014s 00:12:34.051 sys 0m26.308s 00:12:34.051 17:58:26 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@1124 -- # xtrace_disable 00:12:34.051 17:58:26 nvmf_tcp.nvmf_vfio_user_fuzz -- common/autotest_common.sh@10 -- # set +x 00:12:34.051 ************************************ 00:12:34.051 END TEST nvmf_vfio_user_fuzz 00:12:34.051 ************************************ 00:12:34.051 17:58:26 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:12:34.051 17:58:26 nvmf_tcp -- nvmf/nvmf.sh@47 -- # run_test nvmf_host_management /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh --transport=tcp 00:12:34.051 17:58:26 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:12:34.051 17:58:26 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:12:34.051 17:58:26 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:12:34.051 ************************************ 00:12:34.051 START TEST nvmf_host_management 00:12:34.051 ************************************ 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh --transport=tcp 00:12:34.051 * Looking for test storage... 00:12:34.051 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- target/host_management.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@7 -- # uname -s 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:12:34.051 17:58:26 nvmf_tcp.nvmf_host_management -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- paths/export.sh@5 -- # export PATH 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@47 -- # : 0 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@51 -- # have_pci_nics=0 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- target/host_management.sh@11 -- # MALLOC_BDEV_SIZE=64 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- target/host_management.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- target/host_management.sh@105 -- # nvmftestinit 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@448 -- # prepare_net_devs 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@410 -- # local -g is_hw=no 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@412 -- # remove_spdk_ns 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@285 -- # xtrace_disable 00:12:34.052 17:58:26 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@291 -- # pci_devs=() 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@291 -- # local -a pci_devs 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@292 -- # pci_net_devs=() 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@293 -- # pci_drivers=() 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@293 -- # local -A pci_drivers 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@295 -- # net_devs=() 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@295 -- # local -ga net_devs 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@296 -- # e810=() 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@296 -- # local -ga e810 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@297 -- # x722=() 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@297 -- # local -ga x722 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@298 -- # mlx=() 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@298 -- # local -ga mlx 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:12:38.255 Found 0000:86:00.0 (0x8086 - 0x159b) 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:12:38.255 Found 0000:86:00.1 (0x8086 - 0x159b) 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@390 -- # [[ up == up ]] 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:12:38.255 Found net devices under 0000:86:00.0: cvl_0_0 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:12:38.255 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@390 -- # [[ up == up ]] 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:12:38.256 Found net devices under 0000:86:00.1: cvl_0_1 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@414 -- # is_hw=yes 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:12:38.256 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:12:38.256 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.191 ms 00:12:38.256 00:12:38.256 --- 10.0.0.2 ping statistics --- 00:12:38.256 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:38.256 rtt min/avg/max/mdev = 0.191/0.191/0.191/0.000 ms 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:12:38.256 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:12:38.256 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.221 ms 00:12:38.256 00:12:38.256 --- 10.0.0.1 ping statistics --- 00:12:38.256 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:38.256 rtt min/avg/max/mdev = 0.221/0.221/0.221/0.000 ms 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@422 -- # return 0 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- target/host_management.sh@107 -- # nvmf_host_management 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- target/host_management.sh@69 -- # starttarget 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- target/host_management.sh@16 -- # nvmfappstart -m 0x1E 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@722 -- # xtrace_disable 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@481 -- # nvmfpid=527941 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@482 -- # waitforlisten 527941 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@829 -- # '[' -z 527941 ']' 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@834 -- # local max_retries=100 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:38.256 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@838 -- # xtrace_disable 00:12:38.256 17:58:31 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:12:38.256 [2024-07-15 17:58:31.697363] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:12:38.256 [2024-07-15 17:58:31.697405] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:38.256 EAL: No free 2048 kB hugepages reported on node 1 00:12:38.256 [2024-07-15 17:58:31.753286] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:12:38.256 [2024-07-15 17:58:31.834078] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:12:38.256 [2024-07-15 17:58:31.834113] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:12:38.256 [2024-07-15 17:58:31.834120] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:12:38.256 [2024-07-15 17:58:31.834126] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:12:38.256 [2024-07-15 17:58:31.834131] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:12:38.256 [2024-07-15 17:58:31.834253] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:12:38.256 [2024-07-15 17:58:31.834337] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:12:38.256 [2024-07-15 17:58:31.834444] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:12:38.256 [2024-07-15 17:58:31.834445] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 4 00:12:38.824 17:58:32 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:12:38.824 17:58:32 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@862 -- # return 0 00:12:38.824 17:58:32 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:12:38.824 17:58:32 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@728 -- # xtrace_disable 00:12:38.824 17:58:32 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:12:38.824 17:58:32 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:12:38.824 17:58:32 nvmf_tcp.nvmf_host_management -- target/host_management.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:12:38.824 17:58:32 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@559 -- # xtrace_disable 00:12:38.824 17:58:32 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:12:38.824 [2024-07-15 17:58:32.549141] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- target/host_management.sh@20 -- # timing_enter create_subsystem 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@722 -- # xtrace_disable 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- target/host_management.sh@22 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- target/host_management.sh@23 -- # cat 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- target/host_management.sh@30 -- # rpc_cmd 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@559 -- # xtrace_disable 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:12:39.083 Malloc0 00:12:39.083 [2024-07-15 17:58:32.608675] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- target/host_management.sh@31 -- # timing_exit create_subsystems 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@728 -- # xtrace_disable 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- target/host_management.sh@73 -- # perfpid=528208 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- target/host_management.sh@74 -- # waitforlisten 528208 /var/tmp/bdevperf.sock 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@829 -- # '[' -z 528208 ']' 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- target/host_management.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- target/host_management.sh@72 -- # gen_nvmf_target_json 0 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@834 -- # local max_retries=100 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:12:39.083 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@532 -- # config=() 00:12:39.083 17:58:32 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@838 -- # xtrace_disable 00:12:39.084 17:58:32 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@532 -- # local subsystem config 00:12:39.084 17:58:32 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:12:39.084 17:58:32 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:12:39.084 17:58:32 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:12:39.084 { 00:12:39.084 "params": { 00:12:39.084 "name": "Nvme$subsystem", 00:12:39.084 "trtype": "$TEST_TRANSPORT", 00:12:39.084 "traddr": "$NVMF_FIRST_TARGET_IP", 00:12:39.084 "adrfam": "ipv4", 00:12:39.084 "trsvcid": "$NVMF_PORT", 00:12:39.084 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:12:39.084 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:12:39.084 "hdgst": ${hdgst:-false}, 00:12:39.084 "ddgst": ${ddgst:-false} 00:12:39.084 }, 00:12:39.084 "method": "bdev_nvme_attach_controller" 00:12:39.084 } 00:12:39.084 EOF 00:12:39.084 )") 00:12:39.084 17:58:32 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@554 -- # cat 00:12:39.084 17:58:32 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@556 -- # jq . 00:12:39.084 17:58:32 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@557 -- # IFS=, 00:12:39.084 17:58:32 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:12:39.084 "params": { 00:12:39.084 "name": "Nvme0", 00:12:39.084 "trtype": "tcp", 00:12:39.084 "traddr": "10.0.0.2", 00:12:39.084 "adrfam": "ipv4", 00:12:39.084 "trsvcid": "4420", 00:12:39.084 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:12:39.084 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:12:39.084 "hdgst": false, 00:12:39.084 "ddgst": false 00:12:39.084 }, 00:12:39.084 "method": "bdev_nvme_attach_controller" 00:12:39.084 }' 00:12:39.084 [2024-07-15 17:58:32.700809] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:12:39.084 [2024-07-15 17:58:32.700855] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid528208 ] 00:12:39.084 EAL: No free 2048 kB hugepages reported on node 1 00:12:39.084 [2024-07-15 17:58:32.754946] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:39.343 [2024-07-15 17:58:32.830066] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:12:39.343 Running I/O for 10 seconds... 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@862 -- # return 0 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- target/host_management.sh@75 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@559 -- # xtrace_disable 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- target/host_management.sh@78 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill -9 $perfpid || true; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- target/host_management.sh@80 -- # waitforio /var/tmp/bdevperf.sock Nvme0n1 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- target/host_management.sh@45 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- target/host_management.sh@49 -- # '[' -z Nvme0n1 ']' 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- target/host_management.sh@52 -- # local ret=1 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- target/host_management.sh@53 -- # local i 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- target/host_management.sh@54 -- # (( i = 10 )) 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- target/host_management.sh@54 -- # (( i != 0 )) 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- target/host_management.sh@55 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme0n1 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- target/host_management.sh@55 -- # jq -r '.bdevs[0].num_read_ops' 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@559 -- # xtrace_disable 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- target/host_management.sh@55 -- # read_io_count=899 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- target/host_management.sh@58 -- # '[' 899 -ge 100 ']' 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- target/host_management.sh@59 -- # ret=0 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- target/host_management.sh@60 -- # break 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- target/host_management.sh@64 -- # return 0 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- target/host_management.sh@84 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host0 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@559 -- # xtrace_disable 00:12:39.921 17:58:33 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:12:39.922 [2024-07-15 17:58:33.592016] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x242c0a0 is same with the state(5) to be set 00:12:39.922 [2024-07-15 17:58:33.592061] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x242c0a0 is same with the state(5) to be set 00:12:39.922 [2024-07-15 17:58:33.592068] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x242c0a0 is same with the state(5) to be set 00:12:39.922 [2024-07-15 17:58:33.592074] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x242c0a0 is same with the state(5) to be set 00:12:39.922 [2024-07-15 17:58:33.592080] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x242c0a0 is same with the state(5) to be set 00:12:39.922 [2024-07-15 17:58:33.592086] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x242c0a0 is same with the state(5) to be set 00:12:39.922 [2024-07-15 17:58:33.592092] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x242c0a0 is same with the state(5) to be set 00:12:39.922 [2024-07-15 17:58:33.592097] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x242c0a0 is same with the state(5) to be set 00:12:39.922 [2024-07-15 17:58:33.592103] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x242c0a0 is same with the state(5) to be set 00:12:39.922 [2024-07-15 17:58:33.592109] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x242c0a0 is same with the state(5) to be set 00:12:39.922 [2024-07-15 17:58:33.592114] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x242c0a0 is same with the state(5) to be set 00:12:39.922 [2024-07-15 17:58:33.592120] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x242c0a0 is same with the state(5) to be set 00:12:39.922 17:58:33 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:12:39.922 17:58:33 nvmf_tcp.nvmf_host_management -- target/host_management.sh@85 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host0 00:12:39.922 17:58:33 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@559 -- # xtrace_disable 00:12:39.922 17:58:33 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:12:39.922 [2024-07-15 17:58:33.604716] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:12:39.922 [2024-07-15 17:58:33.604747] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.604763] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:12:39.922 [2024-07-15 17:58:33.604771] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.604779] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:12:39.922 [2024-07-15 17:58:33.604787] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.604794] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:12:39.922 [2024-07-15 17:58:33.604801] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.604808] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x179a980 is same with the state(5) to be set 00:12:39.922 [2024-07-15 17:58:33.605532] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605564] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605581] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605603] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605612] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605619] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605628] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605635] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605644] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:768 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605651] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605659] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:896 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605666] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605675] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:1024 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605694] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:1152 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605702] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:1280 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605718] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605726] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:1408 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605733] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605742] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:1536 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:1664 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605765] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605773] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:1792 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605781] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1920 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605796] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:2048 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605812] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:2176 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605827] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:2304 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605852] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:2432 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605860] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605869] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:2560 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605876] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:2688 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605892] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:2816 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605908] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:2944 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605923] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:3072 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605939] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605947] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:3200 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605955] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605963] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:3328 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.922 [2024-07-15 17:58:33.605979] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:3456 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.922 [2024-07-15 17:58:33.605986] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.605994] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:3584 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606010] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:3712 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606025] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:3840 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606033] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606041] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:3968 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606048] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606057] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:4096 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606064] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606073] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:4224 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606079] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:4352 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606096] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606105] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:4480 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606112] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606121] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:4608 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606127] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:4736 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606143] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606151] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:4864 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606159] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606167] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:4992 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606174] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606184] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:5120 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606191] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606200] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:5248 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606206] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606215] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:5376 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606222] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606236] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:5504 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606243] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:5632 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606267] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:5760 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606274] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606282] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:5888 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606289] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606300] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:6016 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606307] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606317] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:6144 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606324] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606332] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:6272 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606340] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606348] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:6400 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606355] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606364] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:6528 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606380] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:6656 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606386] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:6784 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606410] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:6912 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606417] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606425] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:7040 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606434] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606443] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:7168 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606450] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606458] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:7296 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606465] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606473] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:7424 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:7552 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606506] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:7680 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606513] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606521] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:7808 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606528] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606537] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:7936 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606552] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:8064 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:12:39.923 [2024-07-15 17:58:33.606559] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:12:39.923 [2024-07-15 17:58:33.606619] bdev_nvme.c:1612:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x1babb20 was disconnected and freed. reset controller. 00:12:39.923 [2024-07-15 17:58:33.607509] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:12:39.923 17:58:33 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:12:39.923 17:58:33 nvmf_tcp.nvmf_host_management -- target/host_management.sh@87 -- # sleep 1 00:12:39.923 task offset: 0 on job bdev=Nvme0n1 fails 00:12:39.923 00:12:39.923 Latency(us) 00:12:39.923 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:39.924 Job: Nvme0n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:12:39.924 Job: Nvme0n1 ended in about 0.58 seconds with error 00:12:39.924 Verification LBA range: start 0x0 length 0x400 00:12:39.924 Nvme0n1 : 0.58 1757.97 109.87 109.87 0.00 33561.51 1403.33 28835.84 00:12:39.924 =================================================================================================================== 00:12:39.924 Total : 1757.97 109.87 109.87 0.00 33561.51 1403.33 28835.84 00:12:39.924 [2024-07-15 17:58:33.609111] app.c:1053:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:12:39.924 [2024-07-15 17:58:33.609126] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x179a980 (9): Bad file descriptor 00:12:39.924 [2024-07-15 17:58:33.613779] bdev_nvme.c:2067:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:12:41.301 17:58:34 nvmf_tcp.nvmf_host_management -- target/host_management.sh@91 -- # kill -9 528208 00:12:41.301 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/host_management.sh: line 91: kill: (528208) - No such process 00:12:41.301 17:58:34 nvmf_tcp.nvmf_host_management -- target/host_management.sh@91 -- # true 00:12:41.301 17:58:34 nvmf_tcp.nvmf_host_management -- target/host_management.sh@97 -- # rm -f /var/tmp/spdk_cpu_lock_001 /var/tmp/spdk_cpu_lock_002 /var/tmp/spdk_cpu_lock_003 /var/tmp/spdk_cpu_lock_004 00:12:41.301 17:58:34 nvmf_tcp.nvmf_host_management -- target/host_management.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 64 -o 65536 -w verify -t 1 00:12:41.301 17:58:34 nvmf_tcp.nvmf_host_management -- target/host_management.sh@100 -- # gen_nvmf_target_json 0 00:12:41.301 17:58:34 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@532 -- # config=() 00:12:41.301 17:58:34 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@532 -- # local subsystem config 00:12:41.301 17:58:34 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:12:41.301 17:58:34 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:12:41.301 { 00:12:41.301 "params": { 00:12:41.301 "name": "Nvme$subsystem", 00:12:41.301 "trtype": "$TEST_TRANSPORT", 00:12:41.301 "traddr": "$NVMF_FIRST_TARGET_IP", 00:12:41.301 "adrfam": "ipv4", 00:12:41.301 "trsvcid": "$NVMF_PORT", 00:12:41.301 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:12:41.301 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:12:41.301 "hdgst": ${hdgst:-false}, 00:12:41.301 "ddgst": ${ddgst:-false} 00:12:41.301 }, 00:12:41.301 "method": "bdev_nvme_attach_controller" 00:12:41.301 } 00:12:41.301 EOF 00:12:41.301 )") 00:12:41.301 17:58:34 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@554 -- # cat 00:12:41.301 17:58:34 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@556 -- # jq . 00:12:41.301 17:58:34 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@557 -- # IFS=, 00:12:41.301 17:58:34 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:12:41.301 "params": { 00:12:41.301 "name": "Nvme0", 00:12:41.301 "trtype": "tcp", 00:12:41.301 "traddr": "10.0.0.2", 00:12:41.301 "adrfam": "ipv4", 00:12:41.301 "trsvcid": "4420", 00:12:41.301 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:12:41.301 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:12:41.301 "hdgst": false, 00:12:41.301 "ddgst": false 00:12:41.301 }, 00:12:41.301 "method": "bdev_nvme_attach_controller" 00:12:41.301 }' 00:12:41.301 [2024-07-15 17:58:34.660537] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:12:41.301 [2024-07-15 17:58:34.660584] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid528458 ] 00:12:41.301 EAL: No free 2048 kB hugepages reported on node 1 00:12:41.301 [2024-07-15 17:58:34.713980] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:12:41.301 [2024-07-15 17:58:34.784667] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:12:41.301 Running I/O for 1 seconds... 00:12:42.679 00:12:42.679 Latency(us) 00:12:42.679 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:12:42.679 Job: Nvme0n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:12:42.679 Verification LBA range: start 0x0 length 0x400 00:12:42.679 Nvme0n1 : 1.01 1766.75 110.42 0.00 0.00 35683.78 6810.05 35332.45 00:12:42.679 =================================================================================================================== 00:12:42.679 Total : 1766.75 110.42 0.00 0.00 35683.78 6810.05 35332.45 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- target/host_management.sh@102 -- # stoptarget 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- target/host_management.sh@36 -- # rm -f ./local-job0-0-verify.state 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- target/host_management.sh@37 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- target/host_management.sh@38 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- target/host_management.sh@40 -- # nvmftestfini 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@488 -- # nvmfcleanup 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@117 -- # sync 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@120 -- # set +e 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@121 -- # for i in {1..20} 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:12:42.679 rmmod nvme_tcp 00:12:42.679 rmmod nvme_fabrics 00:12:42.679 rmmod nvme_keyring 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@124 -- # set -e 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@125 -- # return 0 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@489 -- # '[' -n 527941 ']' 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@490 -- # killprocess 527941 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@948 -- # '[' -z 527941 ']' 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@952 -- # kill -0 527941 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@953 -- # uname 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 527941 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@966 -- # echo 'killing process with pid 527941' 00:12:42.679 killing process with pid 527941 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@967 -- # kill 527941 00:12:42.679 17:58:36 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@972 -- # wait 527941 00:12:42.938 [2024-07-15 17:58:36.517457] app.c: 711:unclaim_cpu_cores: *ERROR*: Failed to unlink lock fd for core 1, errno: 2 00:12:42.938 17:58:36 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:12:42.938 17:58:36 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:12:42.938 17:58:36 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:12:42.938 17:58:36 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:12:42.938 17:58:36 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@278 -- # remove_spdk_ns 00:12:42.938 17:58:36 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:42.938 17:58:36 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:12:42.938 17:58:36 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:45.476 17:58:38 nvmf_tcp.nvmf_host_management -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:12:45.476 17:58:38 nvmf_tcp.nvmf_host_management -- target/host_management.sh@109 -- # trap - SIGINT SIGTERM EXIT 00:12:45.476 00:12:45.476 real 0m12.425s 00:12:45.476 user 0m22.763s 00:12:45.476 sys 0m5.131s 00:12:45.476 17:58:38 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@1124 -- # xtrace_disable 00:12:45.476 17:58:38 nvmf_tcp.nvmf_host_management -- common/autotest_common.sh@10 -- # set +x 00:12:45.476 ************************************ 00:12:45.476 END TEST nvmf_host_management 00:12:45.476 ************************************ 00:12:45.476 17:58:38 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:12:45.476 17:58:38 nvmf_tcp -- nvmf/nvmf.sh@48 -- # run_test nvmf_lvol /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvol.sh --transport=tcp 00:12:45.476 17:58:38 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:12:45.476 17:58:38 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:12:45.476 17:58:38 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:12:45.476 ************************************ 00:12:45.476 START TEST nvmf_lvol 00:12:45.476 ************************************ 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvol.sh --transport=tcp 00:12:45.476 * Looking for test storage... 00:12:45.476 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@7 -- # uname -s 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- paths/export.sh@5 -- # export PATH 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@47 -- # : 0 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@51 -- # have_pci_nics=0 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@11 -- # MALLOC_BDEV_SIZE=64 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@13 -- # LVOL_BDEV_INIT_SIZE=20 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@14 -- # LVOL_BDEV_FINAL_SIZE=30 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@16 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@18 -- # nvmftestinit 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@448 -- # prepare_net_devs 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@410 -- # local -g is_hw=no 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@412 -- # remove_spdk_ns 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@285 -- # xtrace_disable 00:12:45.476 17:58:38 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@291 -- # pci_devs=() 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@291 -- # local -a pci_devs 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@292 -- # pci_net_devs=() 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@293 -- # pci_drivers=() 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@293 -- # local -A pci_drivers 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@295 -- # net_devs=() 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@295 -- # local -ga net_devs 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@296 -- # e810=() 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@296 -- # local -ga e810 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@297 -- # x722=() 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@297 -- # local -ga x722 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@298 -- # mlx=() 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@298 -- # local -ga mlx 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:12:50.818 Found 0000:86:00.0 (0x8086 - 0x159b) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:12:50.818 Found 0000:86:00.1 (0x8086 - 0x159b) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@390 -- # [[ up == up ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:12:50.818 Found net devices under 0000:86:00.0: cvl_0_0 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@390 -- # [[ up == up ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:12:50.818 Found net devices under 0000:86:00.1: cvl_0_1 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@414 -- # is_hw=yes 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:12:50.818 17:58:43 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:12:50.818 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:12:50.818 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.185 ms 00:12:50.818 00:12:50.818 --- 10.0.0.2 ping statistics --- 00:12:50.818 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:50.818 rtt min/avg/max/mdev = 0.185/0.185/0.185/0.000 ms 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:12:50.818 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:12:50.818 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.181 ms 00:12:50.818 00:12:50.818 --- 10.0.0.1 ping statistics --- 00:12:50.818 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:12:50.818 rtt min/avg/max/mdev = 0.181/0.181/0.181/0.000 ms 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@422 -- # return 0 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@19 -- # nvmfappstart -m 0x7 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@722 -- # xtrace_disable 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@481 -- # nvmfpid=532212 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@482 -- # waitforlisten 532212 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x7 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@829 -- # '[' -z 532212 ']' 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@834 -- # local max_retries=100 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:12:50.818 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@838 -- # xtrace_disable 00:12:50.818 17:58:44 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:12:50.818 [2024-07-15 17:58:44.251103] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:12:50.818 [2024-07-15 17:58:44.251143] [ DPDK EAL parameters: nvmf -c 0x7 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:12:50.818 EAL: No free 2048 kB hugepages reported on node 1 00:12:50.818 [2024-07-15 17:58:44.307211] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:12:50.818 [2024-07-15 17:58:44.387206] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:12:50.818 [2024-07-15 17:58:44.387241] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:12:50.818 [2024-07-15 17:58:44.387248] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:12:50.819 [2024-07-15 17:58:44.387254] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:12:50.819 [2024-07-15 17:58:44.387259] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:12:50.819 [2024-07-15 17:58:44.387317] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:12:50.819 [2024-07-15 17:58:44.387411] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:12:50.819 [2024-07-15 17:58:44.387413] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:12:51.386 17:58:45 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:12:51.386 17:58:45 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@862 -- # return 0 00:12:51.386 17:58:45 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:12:51.386 17:58:45 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@728 -- # xtrace_disable 00:12:51.386 17:58:45 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:12:51.386 17:58:45 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:12:51.386 17:58:45 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:12:51.646 [2024-07-15 17:58:45.260759] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:12:51.646 17:58:45 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:12:51.905 17:58:45 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@24 -- # base_bdevs='Malloc0 ' 00:12:51.905 17:58:45 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:12:52.164 17:58:45 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@25 -- # base_bdevs+=Malloc1 00:12:52.164 17:58:45 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n raid0 -z 64 -r 0 -b 'Malloc0 Malloc1' 00:12:52.164 17:58:45 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore raid0 lvs 00:12:52.423 17:58:46 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@29 -- # lvs=71e13236-c010-4034-97fd-ac972228899b 00:12:52.423 17:58:46 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u 71e13236-c010-4034-97fd-ac972228899b lvol 20 00:12:52.682 17:58:46 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@32 -- # lvol=66618741-3c3d-4e72-a22f-1557fc87f93e 00:12:52.682 17:58:46 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:12:52.682 17:58:46 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 66618741-3c3d-4e72-a22f-1557fc87f93e 00:12:52.941 17:58:46 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:12:53.200 [2024-07-15 17:58:46.740169] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:12:53.200 17:58:46 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:12:53.458 17:58:46 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@42 -- # perf_pid=532709 00:12:53.458 17:58:46 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -o 4096 -q 128 -s 512 -w randwrite -t 10 -c 0x18 00:12:53.458 17:58:46 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@44 -- # sleep 1 00:12:53.458 EAL: No free 2048 kB hugepages reported on node 1 00:12:54.395 17:58:47 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_snapshot 66618741-3c3d-4e72-a22f-1557fc87f93e MY_SNAPSHOT 00:12:54.653 17:58:48 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@47 -- # snapshot=f88f30e4-0e1c-4a59-963b-63d40e29ed6f 00:12:54.653 17:58:48 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_resize 66618741-3c3d-4e72-a22f-1557fc87f93e 30 00:12:54.912 17:58:48 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@49 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_clone f88f30e4-0e1c-4a59-963b-63d40e29ed6f MY_CLONE 00:12:54.912 17:58:48 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@49 -- # clone=13ca7441-4249-495e-a5c1-d5d5ba39334d 00:12:54.912 17:58:48 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_inflate 13ca7441-4249-495e-a5c1-d5d5ba39334d 00:12:55.479 17:58:49 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@53 -- # wait 532709 00:13:05.451 Initializing NVMe Controllers 00:13:05.451 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode0 00:13:05.451 Controller IO queue size 128, less than required. 00:13:05.451 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:13:05.451 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 3 00:13:05.451 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 with lcore 4 00:13:05.451 Initialization complete. Launching workers. 00:13:05.451 ======================================================== 00:13:05.451 Latency(us) 00:13:05.451 Device Information : IOPS MiB/s Average min max 00:13:05.451 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 from core 3: 12516.40 48.89 10231.61 1836.31 59854.77 00:13:05.451 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode0) NSID 1 from core 4: 12428.40 48.55 10301.80 3666.17 49573.11 00:13:05.451 ======================================================== 00:13:05.451 Total : 24944.80 97.44 10266.58 1836.31 59854.77 00:13:05.451 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete 66618741-3c3d-4e72-a22f-1557fc87f93e 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u 71e13236-c010-4034-97fd-ac972228899b 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@60 -- # rm -f 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@62 -- # trap - SIGINT SIGTERM EXIT 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- target/nvmf_lvol.sh@64 -- # nvmftestfini 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@488 -- # nvmfcleanup 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@117 -- # sync 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@120 -- # set +e 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@121 -- # for i in {1..20} 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:13:05.451 rmmod nvme_tcp 00:13:05.451 rmmod nvme_fabrics 00:13:05.451 rmmod nvme_keyring 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@124 -- # set -e 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@125 -- # return 0 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@489 -- # '[' -n 532212 ']' 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@490 -- # killprocess 532212 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@948 -- # '[' -z 532212 ']' 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@952 -- # kill -0 532212 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@953 -- # uname 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:13:05.451 17:58:57 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 532212 00:13:05.451 17:58:58 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:13:05.451 17:58:58 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:13:05.451 17:58:58 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@966 -- # echo 'killing process with pid 532212' 00:13:05.451 killing process with pid 532212 00:13:05.451 17:58:58 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@967 -- # kill 532212 00:13:05.451 17:58:58 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@972 -- # wait 532212 00:13:05.451 17:58:58 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:13:05.451 17:58:58 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:13:05.451 17:58:58 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:13:05.451 17:58:58 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:13:05.451 17:58:58 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@278 -- # remove_spdk_ns 00:13:05.451 17:58:58 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:05.451 17:58:58 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:13:05.451 17:58:58 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:06.825 17:59:00 nvmf_tcp.nvmf_lvol -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:13:06.825 00:13:06.825 real 0m21.598s 00:13:06.825 user 1m3.886s 00:13:06.825 sys 0m6.913s 00:13:06.825 17:59:00 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@1124 -- # xtrace_disable 00:13:06.825 17:59:00 nvmf_tcp.nvmf_lvol -- common/autotest_common.sh@10 -- # set +x 00:13:06.825 ************************************ 00:13:06.825 END TEST nvmf_lvol 00:13:06.825 ************************************ 00:13:06.825 17:59:00 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:13:06.825 17:59:00 nvmf_tcp -- nvmf/nvmf.sh@49 -- # run_test nvmf_lvs_grow /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh --transport=tcp 00:13:06.825 17:59:00 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:13:06.825 17:59:00 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:13:06.825 17:59:00 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:13:06.825 ************************************ 00:13:06.825 START TEST nvmf_lvs_grow 00:13:06.825 ************************************ 00:13:06.825 17:59:00 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh --transport=tcp 00:13:06.825 * Looking for test storage... 00:13:06.825 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:13:06.825 17:59:00 nvmf_tcp.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:13:06.825 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@7 -- # uname -s 00:13:06.825 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:13:06.825 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:13:06.825 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:13:06.825 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:13:06.825 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:13:06.825 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:13:06.825 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- paths/export.sh@5 -- # export PATH 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@47 -- # : 0 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@51 -- # have_pci_nics=0 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@11 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@12 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@98 -- # nvmftestinit 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@448 -- # prepare_net_devs 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@410 -- # local -g is_hw=no 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@412 -- # remove_spdk_ns 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@285 -- # xtrace_disable 00:13:06.826 17:59:00 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@291 -- # pci_devs=() 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@291 -- # local -a pci_devs 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@292 -- # pci_net_devs=() 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@293 -- # pci_drivers=() 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@293 -- # local -A pci_drivers 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@295 -- # net_devs=() 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@295 -- # local -ga net_devs 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@296 -- # e810=() 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@296 -- # local -ga e810 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@297 -- # x722=() 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@297 -- # local -ga x722 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@298 -- # mlx=() 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@298 -- # local -ga mlx 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:13:12.102 Found 0000:86:00.0 (0x8086 - 0x159b) 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:13:12.102 Found 0000:86:00.1 (0x8086 - 0x159b) 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@390 -- # [[ up == up ]] 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:13:12.102 Found net devices under 0000:86:00.0: cvl_0_0 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:12.102 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@390 -- # [[ up == up ]] 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:13:12.103 Found net devices under 0000:86:00.1: cvl_0_1 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@414 -- # is_hw=yes 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:13:12.103 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:13:12.103 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.297 ms 00:13:12.103 00:13:12.103 --- 10.0.0.2 ping statistics --- 00:13:12.103 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:12.103 rtt min/avg/max/mdev = 0.297/0.297/0.297/0.000 ms 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:13:12.103 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:13:12.103 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.257 ms 00:13:12.103 00:13:12.103 --- 10.0.0.1 ping statistics --- 00:13:12.103 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:12.103 rtt min/avg/max/mdev = 0.257/0.257/0.257/0.000 ms 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@422 -- # return 0 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@99 -- # nvmfappstart -m 0x1 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@722 -- # xtrace_disable 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@481 -- # nvmfpid=537965 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@482 -- # waitforlisten 537965 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@829 -- # '[' -z 537965 ']' 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@834 -- # local max_retries=100 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:12.103 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@838 -- # xtrace_disable 00:13:12.103 17:59:05 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:13:12.103 [2024-07-15 17:59:05.783751] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:13:12.103 [2024-07-15 17:59:05.783801] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:12.103 EAL: No free 2048 kB hugepages reported on node 1 00:13:12.362 [2024-07-15 17:59:05.840509] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:12.362 [2024-07-15 17:59:05.919317] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:13:12.362 [2024-07-15 17:59:05.919350] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:13:12.362 [2024-07-15 17:59:05.919356] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:13:12.362 [2024-07-15 17:59:05.919363] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:13:12.362 [2024-07-15 17:59:05.919368] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:13:12.362 [2024-07-15 17:59:05.919384] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:13:12.930 17:59:06 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:13:12.930 17:59:06 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@862 -- # return 0 00:13:12.930 17:59:06 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:13:12.930 17:59:06 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@728 -- # xtrace_disable 00:13:12.930 17:59:06 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:13:12.930 17:59:06 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:13:12.930 17:59:06 nvmf_tcp.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:13:13.190 [2024-07-15 17:59:06.770911] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:13:13.190 17:59:06 nvmf_tcp.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@102 -- # run_test lvs_grow_clean lvs_grow 00:13:13.190 17:59:06 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:13:13.190 17:59:06 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@1105 -- # xtrace_disable 00:13:13.190 17:59:06 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:13:13.190 ************************************ 00:13:13.190 START TEST lvs_grow_clean 00:13:13.190 ************************************ 00:13:13.190 17:59:06 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@1123 -- # lvs_grow 00:13:13.190 17:59:06 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@15 -- # local aio_bdev lvs lvol 00:13:13.190 17:59:06 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@16 -- # local data_clusters free_clusters 00:13:13.190 17:59:06 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@17 -- # local bdevperf_pid run_test_pid 00:13:13.190 17:59:06 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@18 -- # local aio_init_size_mb=200 00:13:13.190 17:59:06 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@19 -- # local aio_final_size_mb=400 00:13:13.190 17:59:06 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@20 -- # local lvol_bdev_size_mb=150 00:13:13.190 17:59:06 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@23 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:13:13.190 17:59:06 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@24 -- # truncate -s 200M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:13:13.190 17:59:06 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:13:13.449 17:59:07 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@25 -- # aio_bdev=aio_bdev 00:13:13.449 17:59:07 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --cluster-sz 4194304 --md-pages-per-cluster-ratio 300 aio_bdev lvs 00:13:13.709 17:59:07 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@28 -- # lvs=edc4b486-5de5-4ff1-9582-417f1a7dd606 00:13:13.709 17:59:07 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u edc4b486-5de5-4ff1-9582-417f1a7dd606 00:13:13.709 17:59:07 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # jq -r '.[0].total_data_clusters' 00:13:13.709 17:59:07 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@29 -- # data_clusters=49 00:13:13.709 17:59:07 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@30 -- # (( data_clusters == 49 )) 00:13:13.709 17:59:07 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u edc4b486-5de5-4ff1-9582-417f1a7dd606 lvol 150 00:13:13.968 17:59:07 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@33 -- # lvol=7825a947-c9f4-4d3d-adb7-79fecf6cbf62 00:13:13.968 17:59:07 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@36 -- # truncate -s 400M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:13:13.968 17:59:07 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_rescan aio_bdev 00:13:14.227 [2024-07-15 17:59:07.715839] bdev_aio.c:1030:bdev_aio_rescan: *NOTICE*: AIO device is resized: bdev name /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev, old block count 51200, new block count 102400 00:13:14.227 [2024-07-15 17:59:07.715888] vbdev_lvol.c: 165:vbdev_lvs_base_bdev_event_cb: *NOTICE*: Unsupported bdev event: type 1 00:13:14.227 true 00:13:14.227 17:59:07 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u edc4b486-5de5-4ff1-9582-417f1a7dd606 00:13:14.227 17:59:07 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # jq -r '.[0].total_data_clusters' 00:13:14.227 17:59:07 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@38 -- # (( data_clusters == 49 )) 00:13:14.227 17:59:07 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:13:14.486 17:59:08 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 7825a947-c9f4-4d3d-adb7-79fecf6cbf62 00:13:14.746 17:59:08 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:13:14.746 [2024-07-15 17:59:08.381846] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:14.746 17:59:08 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:13:15.005 17:59:08 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock -m 0x2 -o 4096 -q 128 -w randwrite -t 10 -S 1 -z 00:13:15.005 17:59:08 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@48 -- # bdevperf_pid=538481 00:13:15.005 17:59:08 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@49 -- # trap 'killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:13:15.005 17:59:08 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@50 -- # waitforlisten 538481 /var/tmp/bdevperf.sock 00:13:15.005 17:59:08 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@829 -- # '[' -z 538481 ']' 00:13:15.005 17:59:08 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:13:15.005 17:59:08 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@834 -- # local max_retries=100 00:13:15.006 17:59:08 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:13:15.006 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:13:15.006 17:59:08 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@838 -- # xtrace_disable 00:13:15.006 17:59:08 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@10 -- # set +x 00:13:15.006 [2024-07-15 17:59:08.595979] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:13:15.006 [2024-07-15 17:59:08.596025] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid538481 ] 00:13:15.006 EAL: No free 2048 kB hugepages reported on node 1 00:13:15.006 [2024-07-15 17:59:08.649850] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:15.006 [2024-07-15 17:59:08.728336] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:13:15.942 17:59:09 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:13:15.942 17:59:09 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@862 -- # return 0 00:13:15.942 17:59:09 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 00:13:15.942 Nvme0n1 00:13:16.201 17:59:09 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_get_bdevs -b Nvme0n1 -t 3000 00:13:16.201 [ 00:13:16.201 { 00:13:16.201 "name": "Nvme0n1", 00:13:16.201 "aliases": [ 00:13:16.201 "7825a947-c9f4-4d3d-adb7-79fecf6cbf62" 00:13:16.201 ], 00:13:16.201 "product_name": "NVMe disk", 00:13:16.201 "block_size": 4096, 00:13:16.201 "num_blocks": 38912, 00:13:16.201 "uuid": "7825a947-c9f4-4d3d-adb7-79fecf6cbf62", 00:13:16.201 "assigned_rate_limits": { 00:13:16.201 "rw_ios_per_sec": 0, 00:13:16.201 "rw_mbytes_per_sec": 0, 00:13:16.201 "r_mbytes_per_sec": 0, 00:13:16.201 "w_mbytes_per_sec": 0 00:13:16.201 }, 00:13:16.201 "claimed": false, 00:13:16.201 "zoned": false, 00:13:16.201 "supported_io_types": { 00:13:16.201 "read": true, 00:13:16.201 "write": true, 00:13:16.201 "unmap": true, 00:13:16.201 "flush": true, 00:13:16.201 "reset": true, 00:13:16.201 "nvme_admin": true, 00:13:16.201 "nvme_io": true, 00:13:16.201 "nvme_io_md": false, 00:13:16.201 "write_zeroes": true, 00:13:16.201 "zcopy": false, 00:13:16.201 "get_zone_info": false, 00:13:16.201 "zone_management": false, 00:13:16.201 "zone_append": false, 00:13:16.201 "compare": true, 00:13:16.201 "compare_and_write": true, 00:13:16.201 "abort": true, 00:13:16.201 "seek_hole": false, 00:13:16.201 "seek_data": false, 00:13:16.201 "copy": true, 00:13:16.201 "nvme_iov_md": false 00:13:16.201 }, 00:13:16.201 "memory_domains": [ 00:13:16.201 { 00:13:16.201 "dma_device_id": "system", 00:13:16.201 "dma_device_type": 1 00:13:16.201 } 00:13:16.201 ], 00:13:16.201 "driver_specific": { 00:13:16.201 "nvme": [ 00:13:16.201 { 00:13:16.201 "trid": { 00:13:16.201 "trtype": "TCP", 00:13:16.201 "adrfam": "IPv4", 00:13:16.201 "traddr": "10.0.0.2", 00:13:16.201 "trsvcid": "4420", 00:13:16.201 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:13:16.201 }, 00:13:16.201 "ctrlr_data": { 00:13:16.201 "cntlid": 1, 00:13:16.201 "vendor_id": "0x8086", 00:13:16.201 "model_number": "SPDK bdev Controller", 00:13:16.201 "serial_number": "SPDK0", 00:13:16.201 "firmware_revision": "24.09", 00:13:16.201 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:13:16.201 "oacs": { 00:13:16.201 "security": 0, 00:13:16.201 "format": 0, 00:13:16.201 "firmware": 0, 00:13:16.201 "ns_manage": 0 00:13:16.201 }, 00:13:16.201 "multi_ctrlr": true, 00:13:16.201 "ana_reporting": false 00:13:16.201 }, 00:13:16.201 "vs": { 00:13:16.201 "nvme_version": "1.3" 00:13:16.201 }, 00:13:16.201 "ns_data": { 00:13:16.201 "id": 1, 00:13:16.201 "can_share": true 00:13:16.201 } 00:13:16.201 } 00:13:16.201 ], 00:13:16.201 "mp_policy": "active_passive" 00:13:16.201 } 00:13:16.201 } 00:13:16.201 ] 00:13:16.201 17:59:09 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@56 -- # run_test_pid=538655 00:13:16.201 17:59:09 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@57 -- # sleep 2 00:13:16.201 17:59:09 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:13:16.201 Running I/O for 10 seconds... 00:13:17.579 Latency(us) 00:13:17.579 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:17.579 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:17.579 Nvme0n1 : 1.00 22166.00 86.59 0.00 0.00 0.00 0.00 0.00 00:13:17.579 =================================================================================================================== 00:13:17.579 Total : 22166.00 86.59 0.00 0.00 0.00 0.00 0.00 00:13:17.579 00:13:18.202 17:59:11 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_grow_lvstore -u edc4b486-5de5-4ff1-9582-417f1a7dd606 00:13:18.461 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:18.462 Nvme0n1 : 2.00 22307.00 87.14 0.00 0.00 0.00 0.00 0.00 00:13:18.462 =================================================================================================================== 00:13:18.462 Total : 22307.00 87.14 0.00 0.00 0.00 0.00 0.00 00:13:18.462 00:13:18.462 true 00:13:18.462 17:59:12 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u edc4b486-5de5-4ff1-9582-417f1a7dd606 00:13:18.462 17:59:12 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # jq -r '.[0].total_data_clusters' 00:13:18.721 17:59:12 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@61 -- # data_clusters=99 00:13:18.721 17:59:12 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@62 -- # (( data_clusters == 99 )) 00:13:18.721 17:59:12 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@65 -- # wait 538655 00:13:19.290 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:19.290 Nvme0n1 : 3.00 22343.33 87.28 0.00 0.00 0.00 0.00 0.00 00:13:19.290 =================================================================================================================== 00:13:19.290 Total : 22343.33 87.28 0.00 0.00 0.00 0.00 0.00 00:13:19.290 00:13:20.226 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:20.226 Nvme0n1 : 4.00 22409.50 87.54 0.00 0.00 0.00 0.00 0.00 00:13:20.226 =================================================================================================================== 00:13:20.226 Total : 22409.50 87.54 0.00 0.00 0.00 0.00 0.00 00:13:20.226 00:13:21.604 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:21.604 Nvme0n1 : 5.00 22449.20 87.69 0.00 0.00 0.00 0.00 0.00 00:13:21.604 =================================================================================================================== 00:13:21.604 Total : 22449.20 87.69 0.00 0.00 0.00 0.00 0.00 00:13:21.604 00:13:22.541 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:22.541 Nvme0n1 : 6.00 22479.67 87.81 0.00 0.00 0.00 0.00 0.00 00:13:22.541 =================================================================================================================== 00:13:22.541 Total : 22479.67 87.81 0.00 0.00 0.00 0.00 0.00 00:13:22.541 00:13:23.478 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:23.478 Nvme0n1 : 7.00 22507.14 87.92 0.00 0.00 0.00 0.00 0.00 00:13:23.478 =================================================================================================================== 00:13:23.478 Total : 22507.14 87.92 0.00 0.00 0.00 0.00 0.00 00:13:23.478 00:13:24.413 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:24.413 Nvme0n1 : 8.00 22530.75 88.01 0.00 0.00 0.00 0.00 0.00 00:13:24.413 =================================================================================================================== 00:13:24.413 Total : 22530.75 88.01 0.00 0.00 0.00 0.00 0.00 00:13:24.413 00:13:25.348 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:25.348 Nvme0n1 : 9.00 22549.11 88.08 0.00 0.00 0.00 0.00 0.00 00:13:25.348 =================================================================================================================== 00:13:25.348 Total : 22549.11 88.08 0.00 0.00 0.00 0.00 0.00 00:13:25.348 00:13:26.280 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:26.280 Nvme0n1 : 10.00 22564.60 88.14 0.00 0.00 0.00 0.00 0.00 00:13:26.280 =================================================================================================================== 00:13:26.280 Total : 22564.60 88.14 0.00 0.00 0.00 0.00 0.00 00:13:26.280 00:13:26.280 00:13:26.280 Latency(us) 00:13:26.280 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:26.280 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:26.280 Nvme0n1 : 10.01 22564.50 88.14 0.00 0.00 5668.41 2806.65 8377.21 00:13:26.280 =================================================================================================================== 00:13:26.280 Total : 22564.50 88.14 0.00 0.00 5668.41 2806.65 8377.21 00:13:26.280 0 00:13:26.280 17:59:19 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@66 -- # killprocess 538481 00:13:26.280 17:59:19 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@948 -- # '[' -z 538481 ']' 00:13:26.281 17:59:19 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@952 -- # kill -0 538481 00:13:26.281 17:59:19 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@953 -- # uname 00:13:26.281 17:59:19 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:13:26.281 17:59:19 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 538481 00:13:26.538 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:13:26.538 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:13:26.538 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@966 -- # echo 'killing process with pid 538481' 00:13:26.538 killing process with pid 538481 00:13:26.538 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@967 -- # kill 538481 00:13:26.538 Received shutdown signal, test time was about 10.000000 seconds 00:13:26.538 00:13:26.538 Latency(us) 00:13:26.538 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:26.538 =================================================================================================================== 00:13:26.538 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:13:26.538 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@972 -- # wait 538481 00:13:26.538 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:13:26.795 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:13:27.053 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # jq -r '.[0].free_clusters' 00:13:27.053 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u edc4b486-5de5-4ff1-9582-417f1a7dd606 00:13:27.053 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@70 -- # free_clusters=61 00:13:27.054 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@72 -- # [[ '' == \d\i\r\t\y ]] 00:13:27.054 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:13:27.312 [2024-07-15 17:59:20.914808] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev aio_bdev being removed: closing lvstore lvs 00:13:27.312 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@85 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u edc4b486-5de5-4ff1-9582-417f1a7dd606 00:13:27.312 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@648 -- # local es=0 00:13:27.312 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u edc4b486-5de5-4ff1-9582-417f1a7dd606 00:13:27.312 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:27.312 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:13:27.312 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:27.312 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:13:27.312 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:27.312 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:13:27.312 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:27.312 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:13:27.312 17:59:20 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u edc4b486-5de5-4ff1-9582-417f1a7dd606 00:13:27.571 request: 00:13:27.571 { 00:13:27.571 "uuid": "edc4b486-5de5-4ff1-9582-417f1a7dd606", 00:13:27.571 "method": "bdev_lvol_get_lvstores", 00:13:27.571 "req_id": 1 00:13:27.571 } 00:13:27.571 Got JSON-RPC error response 00:13:27.571 response: 00:13:27.571 { 00:13:27.571 "code": -19, 00:13:27.571 "message": "No such device" 00:13:27.571 } 00:13:27.571 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@651 -- # es=1 00:13:27.571 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:13:27.571 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:13:27.571 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:13:27.571 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:13:27.829 aio_bdev 00:13:27.829 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@87 -- # waitforbdev 7825a947-c9f4-4d3d-adb7-79fecf6cbf62 00:13:27.829 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@897 -- # local bdev_name=7825a947-c9f4-4d3d-adb7-79fecf6cbf62 00:13:27.829 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:13:27.829 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@899 -- # local i 00:13:27.829 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:13:27.829 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:13:27.829 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:13:27.829 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b 7825a947-c9f4-4d3d-adb7-79fecf6cbf62 -t 2000 00:13:28.087 [ 00:13:28.087 { 00:13:28.087 "name": "7825a947-c9f4-4d3d-adb7-79fecf6cbf62", 00:13:28.087 "aliases": [ 00:13:28.087 "lvs/lvol" 00:13:28.087 ], 00:13:28.087 "product_name": "Logical Volume", 00:13:28.087 "block_size": 4096, 00:13:28.087 "num_blocks": 38912, 00:13:28.087 "uuid": "7825a947-c9f4-4d3d-adb7-79fecf6cbf62", 00:13:28.087 "assigned_rate_limits": { 00:13:28.087 "rw_ios_per_sec": 0, 00:13:28.087 "rw_mbytes_per_sec": 0, 00:13:28.087 "r_mbytes_per_sec": 0, 00:13:28.087 "w_mbytes_per_sec": 0 00:13:28.087 }, 00:13:28.087 "claimed": false, 00:13:28.087 "zoned": false, 00:13:28.087 "supported_io_types": { 00:13:28.087 "read": true, 00:13:28.087 "write": true, 00:13:28.087 "unmap": true, 00:13:28.087 "flush": false, 00:13:28.087 "reset": true, 00:13:28.087 "nvme_admin": false, 00:13:28.087 "nvme_io": false, 00:13:28.087 "nvme_io_md": false, 00:13:28.087 "write_zeroes": true, 00:13:28.087 "zcopy": false, 00:13:28.087 "get_zone_info": false, 00:13:28.087 "zone_management": false, 00:13:28.087 "zone_append": false, 00:13:28.087 "compare": false, 00:13:28.087 "compare_and_write": false, 00:13:28.087 "abort": false, 00:13:28.087 "seek_hole": true, 00:13:28.087 "seek_data": true, 00:13:28.087 "copy": false, 00:13:28.087 "nvme_iov_md": false 00:13:28.087 }, 00:13:28.087 "driver_specific": { 00:13:28.087 "lvol": { 00:13:28.087 "lvol_store_uuid": "edc4b486-5de5-4ff1-9582-417f1a7dd606", 00:13:28.087 "base_bdev": "aio_bdev", 00:13:28.087 "thin_provision": false, 00:13:28.087 "num_allocated_clusters": 38, 00:13:28.087 "snapshot": false, 00:13:28.087 "clone": false, 00:13:28.087 "esnap_clone": false 00:13:28.087 } 00:13:28.087 } 00:13:28.087 } 00:13:28.087 ] 00:13:28.087 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@905 -- # return 0 00:13:28.087 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u edc4b486-5de5-4ff1-9582-417f1a7dd606 00:13:28.087 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # jq -r '.[0].free_clusters' 00:13:28.346 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@88 -- # (( free_clusters == 61 )) 00:13:28.346 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u edc4b486-5de5-4ff1-9582-417f1a7dd606 00:13:28.346 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # jq -r '.[0].total_data_clusters' 00:13:28.346 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@89 -- # (( data_clusters == 99 )) 00:13:28.346 17:59:21 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete 7825a947-c9f4-4d3d-adb7-79fecf6cbf62 00:13:28.605 17:59:22 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u edc4b486-5de5-4ff1-9582-417f1a7dd606 00:13:28.864 17:59:22 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:13:28.864 17:59:22 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- target/nvmf_lvs_grow.sh@95 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:13:28.864 00:13:28.864 real 0m15.734s 00:13:28.864 user 0m15.379s 00:13:28.864 sys 0m1.436s 00:13:28.864 17:59:22 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@1124 -- # xtrace_disable 00:13:28.864 17:59:22 nvmf_tcp.nvmf_lvs_grow.lvs_grow_clean -- common/autotest_common.sh@10 -- # set +x 00:13:28.864 ************************************ 00:13:28.864 END TEST lvs_grow_clean 00:13:28.864 ************************************ 00:13:28.864 17:59:22 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@1142 -- # return 0 00:13:28.864 17:59:22 nvmf_tcp.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@103 -- # run_test lvs_grow_dirty lvs_grow dirty 00:13:28.864 17:59:22 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:13:28.864 17:59:22 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@1105 -- # xtrace_disable 00:13:28.864 17:59:22 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:13:29.134 ************************************ 00:13:29.134 START TEST lvs_grow_dirty 00:13:29.134 ************************************ 00:13:29.134 17:59:22 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@1123 -- # lvs_grow dirty 00:13:29.134 17:59:22 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@15 -- # local aio_bdev lvs lvol 00:13:29.134 17:59:22 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@16 -- # local data_clusters free_clusters 00:13:29.134 17:59:22 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@17 -- # local bdevperf_pid run_test_pid 00:13:29.134 17:59:22 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@18 -- # local aio_init_size_mb=200 00:13:29.134 17:59:22 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@19 -- # local aio_final_size_mb=400 00:13:29.134 17:59:22 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@20 -- # local lvol_bdev_size_mb=150 00:13:29.134 17:59:22 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@23 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:13:29.134 17:59:22 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@24 -- # truncate -s 200M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:13:29.134 17:59:22 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:13:29.134 17:59:22 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@25 -- # aio_bdev=aio_bdev 00:13:29.135 17:59:22 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create_lvstore --cluster-sz 4194304 --md-pages-per-cluster-ratio 300 aio_bdev lvs 00:13:29.393 17:59:22 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@28 -- # lvs=c2cf3162-0b71-4374-8533-4b15b0ed84a4 00:13:29.393 17:59:22 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u c2cf3162-0b71-4374-8533-4b15b0ed84a4 00:13:29.393 17:59:22 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # jq -r '.[0].total_data_clusters' 00:13:29.651 17:59:23 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@29 -- # data_clusters=49 00:13:29.651 17:59:23 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@30 -- # (( data_clusters == 49 )) 00:13:29.651 17:59:23 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_create -u c2cf3162-0b71-4374-8533-4b15b0ed84a4 lvol 150 00:13:29.651 17:59:23 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@33 -- # lvol=da5ec191-0efd-4ecd-a5b2-881165b3786e 00:13:29.651 17:59:23 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@36 -- # truncate -s 400M /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:13:29.651 17:59:23 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_rescan aio_bdev 00:13:29.909 [2024-07-15 17:59:23.481906] bdev_aio.c:1030:bdev_aio_rescan: *NOTICE*: AIO device is resized: bdev name /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev, old block count 51200, new block count 102400 00:13:29.909 [2024-07-15 17:59:23.481958] vbdev_lvol.c: 165:vbdev_lvs_base_bdev_event_cb: *NOTICE*: Unsupported bdev event: type 1 00:13:29.909 true 00:13:29.909 17:59:23 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u c2cf3162-0b71-4374-8533-4b15b0ed84a4 00:13:29.909 17:59:23 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # jq -r '.[0].total_data_clusters' 00:13:30.168 17:59:23 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@38 -- # (( data_clusters == 49 )) 00:13:30.168 17:59:23 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a -s SPDK0 00:13:30.168 17:59:23 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 da5ec191-0efd-4ecd-a5b2-881165b3786e 00:13:30.427 17:59:23 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:13:30.685 [2024-07-15 17:59:24.159923] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:30.685 17:59:24 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:13:30.685 17:59:24 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@48 -- # bdevperf_pid=541162 00:13:30.685 17:59:24 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock -m 0x2 -o 4096 -q 128 -w randwrite -t 10 -S 1 -z 00:13:30.685 17:59:24 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@49 -- # trap 'killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:13:30.685 17:59:24 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@50 -- # waitforlisten 541162 /var/tmp/bdevperf.sock 00:13:30.685 17:59:24 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@829 -- # '[' -z 541162 ']' 00:13:30.685 17:59:24 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:13:30.685 17:59:24 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@834 -- # local max_retries=100 00:13:30.685 17:59:24 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:13:30.685 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:13:30.685 17:59:24 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@838 -- # xtrace_disable 00:13:30.685 17:59:24 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:13:30.685 [2024-07-15 17:59:24.377624] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:13:30.685 [2024-07-15 17:59:24.377671] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid541162 ] 00:13:30.685 EAL: No free 2048 kB hugepages reported on node 1 00:13:30.944 [2024-07-15 17:59:24.431219] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:30.944 [2024-07-15 17:59:24.509943] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:13:31.513 17:59:25 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:13:31.513 17:59:25 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@862 -- # return 0 00:13:31.513 17:59:25 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 00:13:32.081 Nvme0n1 00:13:32.081 17:59:25 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_get_bdevs -b Nvme0n1 -t 3000 00:13:32.081 [ 00:13:32.081 { 00:13:32.081 "name": "Nvme0n1", 00:13:32.081 "aliases": [ 00:13:32.081 "da5ec191-0efd-4ecd-a5b2-881165b3786e" 00:13:32.081 ], 00:13:32.081 "product_name": "NVMe disk", 00:13:32.081 "block_size": 4096, 00:13:32.081 "num_blocks": 38912, 00:13:32.081 "uuid": "da5ec191-0efd-4ecd-a5b2-881165b3786e", 00:13:32.081 "assigned_rate_limits": { 00:13:32.081 "rw_ios_per_sec": 0, 00:13:32.081 "rw_mbytes_per_sec": 0, 00:13:32.081 "r_mbytes_per_sec": 0, 00:13:32.081 "w_mbytes_per_sec": 0 00:13:32.081 }, 00:13:32.081 "claimed": false, 00:13:32.081 "zoned": false, 00:13:32.081 "supported_io_types": { 00:13:32.081 "read": true, 00:13:32.081 "write": true, 00:13:32.081 "unmap": true, 00:13:32.081 "flush": true, 00:13:32.081 "reset": true, 00:13:32.081 "nvme_admin": true, 00:13:32.081 "nvme_io": true, 00:13:32.081 "nvme_io_md": false, 00:13:32.081 "write_zeroes": true, 00:13:32.081 "zcopy": false, 00:13:32.081 "get_zone_info": false, 00:13:32.081 "zone_management": false, 00:13:32.081 "zone_append": false, 00:13:32.081 "compare": true, 00:13:32.081 "compare_and_write": true, 00:13:32.081 "abort": true, 00:13:32.081 "seek_hole": false, 00:13:32.081 "seek_data": false, 00:13:32.081 "copy": true, 00:13:32.081 "nvme_iov_md": false 00:13:32.081 }, 00:13:32.081 "memory_domains": [ 00:13:32.081 { 00:13:32.081 "dma_device_id": "system", 00:13:32.081 "dma_device_type": 1 00:13:32.081 } 00:13:32.081 ], 00:13:32.081 "driver_specific": { 00:13:32.081 "nvme": [ 00:13:32.081 { 00:13:32.081 "trid": { 00:13:32.081 "trtype": "TCP", 00:13:32.081 "adrfam": "IPv4", 00:13:32.081 "traddr": "10.0.0.2", 00:13:32.081 "trsvcid": "4420", 00:13:32.081 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:13:32.081 }, 00:13:32.081 "ctrlr_data": { 00:13:32.081 "cntlid": 1, 00:13:32.081 "vendor_id": "0x8086", 00:13:32.081 "model_number": "SPDK bdev Controller", 00:13:32.081 "serial_number": "SPDK0", 00:13:32.081 "firmware_revision": "24.09", 00:13:32.081 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:13:32.081 "oacs": { 00:13:32.081 "security": 0, 00:13:32.081 "format": 0, 00:13:32.081 "firmware": 0, 00:13:32.081 "ns_manage": 0 00:13:32.081 }, 00:13:32.081 "multi_ctrlr": true, 00:13:32.081 "ana_reporting": false 00:13:32.081 }, 00:13:32.081 "vs": { 00:13:32.081 "nvme_version": "1.3" 00:13:32.081 }, 00:13:32.081 "ns_data": { 00:13:32.081 "id": 1, 00:13:32.081 "can_share": true 00:13:32.081 } 00:13:32.081 } 00:13:32.081 ], 00:13:32.081 "mp_policy": "active_passive" 00:13:32.081 } 00:13:32.081 } 00:13:32.081 ] 00:13:32.081 17:59:25 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@56 -- # run_test_pid=541396 00:13:32.081 17:59:25 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@57 -- # sleep 2 00:13:32.081 17:59:25 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:13:32.081 Running I/O for 10 seconds... 00:13:33.458 Latency(us) 00:13:33.458 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:33.458 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:33.458 Nvme0n1 : 1.00 22997.00 89.83 0.00 0.00 0.00 0.00 0.00 00:13:33.458 =================================================================================================================== 00:13:33.458 Total : 22997.00 89.83 0.00 0.00 0.00 0.00 0.00 00:13:33.458 00:13:34.026 17:59:27 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_grow_lvstore -u c2cf3162-0b71-4374-8533-4b15b0ed84a4 00:13:34.285 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:34.285 Nvme0n1 : 2.00 23210.50 90.67 0.00 0.00 0.00 0.00 0.00 00:13:34.285 =================================================================================================================== 00:13:34.285 Total : 23210.50 90.67 0.00 0.00 0.00 0.00 0.00 00:13:34.285 00:13:34.285 true 00:13:34.285 17:59:27 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # jq -r '.[0].total_data_clusters' 00:13:34.285 17:59:27 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u c2cf3162-0b71-4374-8533-4b15b0ed84a4 00:13:34.544 17:59:28 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@61 -- # data_clusters=99 00:13:34.544 17:59:28 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@62 -- # (( data_clusters == 99 )) 00:13:34.544 17:59:28 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@65 -- # wait 541396 00:13:35.115 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:35.115 Nvme0n1 : 3.00 23260.33 90.86 0.00 0.00 0.00 0.00 0.00 00:13:35.115 =================================================================================================================== 00:13:35.115 Total : 23260.33 90.86 0.00 0.00 0.00 0.00 0.00 00:13:35.115 00:13:36.092 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:36.092 Nvme0n1 : 4.00 23282.50 90.95 0.00 0.00 0.00 0.00 0.00 00:13:36.092 =================================================================================================================== 00:13:36.092 Total : 23282.50 90.95 0.00 0.00 0.00 0.00 0.00 00:13:36.092 00:13:37.470 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:37.470 Nvme0n1 : 5.00 23296.60 91.00 0.00 0.00 0.00 0.00 0.00 00:13:37.470 =================================================================================================================== 00:13:37.470 Total : 23296.60 91.00 0.00 0.00 0.00 0.00 0.00 00:13:37.470 00:13:38.404 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:38.404 Nvme0n1 : 6.00 23334.00 91.15 0.00 0.00 0.00 0.00 0.00 00:13:38.404 =================================================================================================================== 00:13:38.404 Total : 23334.00 91.15 0.00 0.00 0.00 0.00 0.00 00:13:38.404 00:13:39.340 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:39.340 Nvme0n1 : 7.00 23353.86 91.23 0.00 0.00 0.00 0.00 0.00 00:13:39.340 =================================================================================================================== 00:13:39.340 Total : 23353.86 91.23 0.00 0.00 0.00 0.00 0.00 00:13:39.340 00:13:40.277 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:40.277 Nvme0n1 : 8.00 23367.38 91.28 0.00 0.00 0.00 0.00 0.00 00:13:40.277 =================================================================================================================== 00:13:40.277 Total : 23367.38 91.28 0.00 0.00 0.00 0.00 0.00 00:13:40.277 00:13:41.215 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:41.215 Nvme0n1 : 9.00 23385.00 91.35 0.00 0.00 0.00 0.00 0.00 00:13:41.215 =================================================================================================================== 00:13:41.215 Total : 23385.00 91.35 0.00 0.00 0.00 0.00 0.00 00:13:41.215 00:13:42.150 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:42.150 Nvme0n1 : 10.00 23406.00 91.43 0.00 0.00 0.00 0.00 0.00 00:13:42.150 =================================================================================================================== 00:13:42.150 Total : 23406.00 91.43 0.00 0.00 0.00 0.00 0.00 00:13:42.150 00:13:42.150 00:13:42.150 Latency(us) 00:13:42.150 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:42.150 Job: Nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:13:42.150 Nvme0n1 : 10.00 23407.59 91.44 0.00 0.00 5465.13 3319.54 13620.09 00:13:42.150 =================================================================================================================== 00:13:42.150 Total : 23407.59 91.44 0.00 0.00 5465.13 3319.54 13620.09 00:13:42.150 0 00:13:42.150 17:59:35 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@66 -- # killprocess 541162 00:13:42.150 17:59:35 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@948 -- # '[' -z 541162 ']' 00:13:42.150 17:59:35 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@952 -- # kill -0 541162 00:13:42.150 17:59:35 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@953 -- # uname 00:13:42.150 17:59:35 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:13:42.150 17:59:35 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 541162 00:13:42.407 17:59:35 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:13:42.408 17:59:35 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:13:42.408 17:59:35 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@966 -- # echo 'killing process with pid 541162' 00:13:42.408 killing process with pid 541162 00:13:42.408 17:59:35 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@967 -- # kill 541162 00:13:42.408 Received shutdown signal, test time was about 10.000000 seconds 00:13:42.408 00:13:42.408 Latency(us) 00:13:42.408 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:42.408 =================================================================================================================== 00:13:42.408 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:13:42.408 17:59:35 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@972 -- # wait 541162 00:13:42.408 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:13:42.665 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@69 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:13:42.978 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u c2cf3162-0b71-4374-8533-4b15b0ed84a4 00:13:42.978 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # jq -r '.[0].free_clusters' 00:13:42.978 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@70 -- # free_clusters=61 00:13:42.978 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@72 -- # [[ dirty == \d\i\r\t\y ]] 00:13:42.978 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@74 -- # kill -9 537965 00:13:42.978 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@75 -- # wait 537965 00:13:42.978 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nvmf_lvs_grow.sh: line 75: 537965 Killed "${NVMF_APP[@]}" "$@" 00:13:42.978 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@75 -- # true 00:13:42.978 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@76 -- # nvmfappstart -m 0x1 00:13:42.978 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:13:42.978 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@722 -- # xtrace_disable 00:13:42.978 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:13:42.978 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@481 -- # nvmfpid=543239 00:13:42.978 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@482 -- # waitforlisten 543239 00:13:42.978 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:13:42.978 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@829 -- # '[' -z 543239 ']' 00:13:42.978 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:42.978 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@834 -- # local max_retries=100 00:13:42.978 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:42.978 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:42.978 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@838 -- # xtrace_disable 00:13:42.978 17:59:36 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:13:42.978 [2024-07-15 17:59:36.694816] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:13:42.978 [2024-07-15 17:59:36.694861] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:43.236 EAL: No free 2048 kB hugepages reported on node 1 00:13:43.236 [2024-07-15 17:59:36.747231] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:43.236 [2024-07-15 17:59:36.825687] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:13:43.236 [2024-07-15 17:59:36.825723] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:13:43.236 [2024-07-15 17:59:36.825730] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:13:43.236 [2024-07-15 17:59:36.825736] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:13:43.236 [2024-07-15 17:59:36.825741] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:13:43.236 [2024-07-15 17:59:36.825759] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:13:43.799 17:59:37 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:13:43.799 17:59:37 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@862 -- # return 0 00:13:43.799 17:59:37 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:13:43.799 17:59:37 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@728 -- # xtrace_disable 00:13:43.799 17:59:37 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:13:44.057 17:59:37 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:13:44.057 17:59:37 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:13:44.057 [2024-07-15 17:59:37.695198] blobstore.c:4865:bs_recover: *NOTICE*: Performing recovery on blobstore 00:13:44.057 [2024-07-15 17:59:37.695306] blobstore.c:4812:bs_load_replay_md_cpl: *NOTICE*: Recover: blob 0x0 00:13:44.057 [2024-07-15 17:59:37.695332] blobstore.c:4812:bs_load_replay_md_cpl: *NOTICE*: Recover: blob 0x1 00:13:44.057 17:59:37 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@77 -- # aio_bdev=aio_bdev 00:13:44.057 17:59:37 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@78 -- # waitforbdev da5ec191-0efd-4ecd-a5b2-881165b3786e 00:13:44.057 17:59:37 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@897 -- # local bdev_name=da5ec191-0efd-4ecd-a5b2-881165b3786e 00:13:44.057 17:59:37 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:13:44.057 17:59:37 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@899 -- # local i 00:13:44.057 17:59:37 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:13:44.057 17:59:37 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:13:44.057 17:59:37 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:13:44.315 17:59:37 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b da5ec191-0efd-4ecd-a5b2-881165b3786e -t 2000 00:13:44.573 [ 00:13:44.573 { 00:13:44.573 "name": "da5ec191-0efd-4ecd-a5b2-881165b3786e", 00:13:44.573 "aliases": [ 00:13:44.573 "lvs/lvol" 00:13:44.573 ], 00:13:44.573 "product_name": "Logical Volume", 00:13:44.573 "block_size": 4096, 00:13:44.573 "num_blocks": 38912, 00:13:44.573 "uuid": "da5ec191-0efd-4ecd-a5b2-881165b3786e", 00:13:44.573 "assigned_rate_limits": { 00:13:44.573 "rw_ios_per_sec": 0, 00:13:44.573 "rw_mbytes_per_sec": 0, 00:13:44.573 "r_mbytes_per_sec": 0, 00:13:44.573 "w_mbytes_per_sec": 0 00:13:44.573 }, 00:13:44.573 "claimed": false, 00:13:44.573 "zoned": false, 00:13:44.573 "supported_io_types": { 00:13:44.573 "read": true, 00:13:44.573 "write": true, 00:13:44.573 "unmap": true, 00:13:44.573 "flush": false, 00:13:44.573 "reset": true, 00:13:44.573 "nvme_admin": false, 00:13:44.573 "nvme_io": false, 00:13:44.573 "nvme_io_md": false, 00:13:44.573 "write_zeroes": true, 00:13:44.573 "zcopy": false, 00:13:44.573 "get_zone_info": false, 00:13:44.573 "zone_management": false, 00:13:44.573 "zone_append": false, 00:13:44.573 "compare": false, 00:13:44.573 "compare_and_write": false, 00:13:44.573 "abort": false, 00:13:44.573 "seek_hole": true, 00:13:44.573 "seek_data": true, 00:13:44.573 "copy": false, 00:13:44.573 "nvme_iov_md": false 00:13:44.573 }, 00:13:44.573 "driver_specific": { 00:13:44.573 "lvol": { 00:13:44.573 "lvol_store_uuid": "c2cf3162-0b71-4374-8533-4b15b0ed84a4", 00:13:44.573 "base_bdev": "aio_bdev", 00:13:44.573 "thin_provision": false, 00:13:44.573 "num_allocated_clusters": 38, 00:13:44.573 "snapshot": false, 00:13:44.573 "clone": false, 00:13:44.573 "esnap_clone": false 00:13:44.573 } 00:13:44.573 } 00:13:44.573 } 00:13:44.573 ] 00:13:44.573 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@905 -- # return 0 00:13:44.573 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u c2cf3162-0b71-4374-8533-4b15b0ed84a4 00:13:44.573 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # jq -r '.[0].free_clusters' 00:13:44.573 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@79 -- # (( free_clusters == 61 )) 00:13:44.573 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u c2cf3162-0b71-4374-8533-4b15b0ed84a4 00:13:44.573 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # jq -r '.[0].total_data_clusters' 00:13:44.832 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@80 -- # (( data_clusters == 99 )) 00:13:44.832 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:13:45.090 [2024-07-15 17:59:38.579741] vbdev_lvol.c: 150:vbdev_lvs_hotremove_cb: *NOTICE*: bdev aio_bdev being removed: closing lvstore lvs 00:13:45.090 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@85 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u c2cf3162-0b71-4374-8533-4b15b0ed84a4 00:13:45.090 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@648 -- # local es=0 00:13:45.090 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u c2cf3162-0b71-4374-8533-4b15b0ed84a4 00:13:45.090 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:45.090 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:13:45.090 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:45.090 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:13:45.090 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:45.090 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:13:45.090 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:13:45.090 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py ]] 00:13:45.090 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u c2cf3162-0b71-4374-8533-4b15b0ed84a4 00:13:45.090 request: 00:13:45.091 { 00:13:45.091 "uuid": "c2cf3162-0b71-4374-8533-4b15b0ed84a4", 00:13:45.091 "method": "bdev_lvol_get_lvstores", 00:13:45.091 "req_id": 1 00:13:45.091 } 00:13:45.091 Got JSON-RPC error response 00:13:45.091 response: 00:13:45.091 { 00:13:45.091 "code": -19, 00:13:45.091 "message": "No such device" 00:13:45.091 } 00:13:45.091 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@651 -- # es=1 00:13:45.091 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:13:45.091 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:13:45.091 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:13:45.091 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@86 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_create /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev aio_bdev 4096 00:13:45.349 aio_bdev 00:13:45.349 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@87 -- # waitforbdev da5ec191-0efd-4ecd-a5b2-881165b3786e 00:13:45.349 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@897 -- # local bdev_name=da5ec191-0efd-4ecd-a5b2-881165b3786e 00:13:45.349 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@898 -- # local bdev_timeout= 00:13:45.349 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@899 -- # local i 00:13:45.349 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@900 -- # [[ -z '' ]] 00:13:45.349 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@900 -- # bdev_timeout=2000 00:13:45.349 17:59:38 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@902 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_wait_for_examine 00:13:45.608 17:59:39 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@904 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_get_bdevs -b da5ec191-0efd-4ecd-a5b2-881165b3786e -t 2000 00:13:45.608 [ 00:13:45.608 { 00:13:45.608 "name": "da5ec191-0efd-4ecd-a5b2-881165b3786e", 00:13:45.608 "aliases": [ 00:13:45.608 "lvs/lvol" 00:13:45.608 ], 00:13:45.608 "product_name": "Logical Volume", 00:13:45.608 "block_size": 4096, 00:13:45.608 "num_blocks": 38912, 00:13:45.608 "uuid": "da5ec191-0efd-4ecd-a5b2-881165b3786e", 00:13:45.608 "assigned_rate_limits": { 00:13:45.608 "rw_ios_per_sec": 0, 00:13:45.608 "rw_mbytes_per_sec": 0, 00:13:45.608 "r_mbytes_per_sec": 0, 00:13:45.608 "w_mbytes_per_sec": 0 00:13:45.608 }, 00:13:45.608 "claimed": false, 00:13:45.608 "zoned": false, 00:13:45.608 "supported_io_types": { 00:13:45.608 "read": true, 00:13:45.608 "write": true, 00:13:45.608 "unmap": true, 00:13:45.608 "flush": false, 00:13:45.608 "reset": true, 00:13:45.608 "nvme_admin": false, 00:13:45.608 "nvme_io": false, 00:13:45.608 "nvme_io_md": false, 00:13:45.608 "write_zeroes": true, 00:13:45.608 "zcopy": false, 00:13:45.608 "get_zone_info": false, 00:13:45.608 "zone_management": false, 00:13:45.608 "zone_append": false, 00:13:45.608 "compare": false, 00:13:45.608 "compare_and_write": false, 00:13:45.608 "abort": false, 00:13:45.608 "seek_hole": true, 00:13:45.608 "seek_data": true, 00:13:45.608 "copy": false, 00:13:45.608 "nvme_iov_md": false 00:13:45.608 }, 00:13:45.608 "driver_specific": { 00:13:45.608 "lvol": { 00:13:45.608 "lvol_store_uuid": "c2cf3162-0b71-4374-8533-4b15b0ed84a4", 00:13:45.608 "base_bdev": "aio_bdev", 00:13:45.608 "thin_provision": false, 00:13:45.608 "num_allocated_clusters": 38, 00:13:45.608 "snapshot": false, 00:13:45.608 "clone": false, 00:13:45.608 "esnap_clone": false 00:13:45.608 } 00:13:45.608 } 00:13:45.608 } 00:13:45.608 ] 00:13:45.608 17:59:39 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@905 -- # return 0 00:13:45.608 17:59:39 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u c2cf3162-0b71-4374-8533-4b15b0ed84a4 00:13:45.608 17:59:39 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # jq -r '.[0].free_clusters' 00:13:45.867 17:59:39 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@88 -- # (( free_clusters == 61 )) 00:13:45.867 17:59:39 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_get_lvstores -u c2cf3162-0b71-4374-8533-4b15b0ed84a4 00:13:45.867 17:59:39 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # jq -r '.[0].total_data_clusters' 00:13:46.125 17:59:39 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@89 -- # (( data_clusters == 99 )) 00:13:46.125 17:59:39 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@92 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete da5ec191-0efd-4ecd-a5b2-881165b3786e 00:13:46.125 17:59:39 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_lvol_delete_lvstore -u c2cf3162-0b71-4374-8533-4b15b0ed84a4 00:13:46.384 17:59:39 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@94 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_aio_delete aio_bdev 00:13:46.642 17:59:40 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- target/nvmf_lvs_grow.sh@95 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/aio_bdev 00:13:46.642 00:13:46.642 real 0m17.569s 00:13:46.642 user 0m44.887s 00:13:46.642 sys 0m3.918s 00:13:46.642 17:59:40 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@1124 -- # xtrace_disable 00:13:46.642 17:59:40 nvmf_tcp.nvmf_lvs_grow.lvs_grow_dirty -- common/autotest_common.sh@10 -- # set +x 00:13:46.642 ************************************ 00:13:46.642 END TEST lvs_grow_dirty 00:13:46.642 ************************************ 00:13:46.642 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@1142 -- # return 0 00:13:46.642 17:59:40 nvmf_tcp.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@1 -- # process_shm --id 0 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@806 -- # type=--id 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@807 -- # id=0 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@808 -- # '[' --id = --pid ']' 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@812 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@812 -- # shm_files=nvmf_trace.0 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@814 -- # [[ -z nvmf_trace.0 ]] 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@818 -- # for n in $shm_files 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@819 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:13:46.643 nvmf_trace.0 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@821 -- # return 0 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- target/nvmf_lvs_grow.sh@1 -- # nvmftestfini 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@488 -- # nvmfcleanup 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@117 -- # sync 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@120 -- # set +e 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@121 -- # for i in {1..20} 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:13:46.643 rmmod nvme_tcp 00:13:46.643 rmmod nvme_fabrics 00:13:46.643 rmmod nvme_keyring 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@124 -- # set -e 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@125 -- # return 0 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@489 -- # '[' -n 543239 ']' 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@490 -- # killprocess 543239 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@948 -- # '[' -z 543239 ']' 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@952 -- # kill -0 543239 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@953 -- # uname 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:13:46.643 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 543239 00:13:46.902 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:13:46.902 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:13:46.902 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@966 -- # echo 'killing process with pid 543239' 00:13:46.902 killing process with pid 543239 00:13:46.902 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@967 -- # kill 543239 00:13:46.902 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@972 -- # wait 543239 00:13:46.902 17:59:40 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:13:46.902 17:59:40 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:13:46.902 17:59:40 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:13:46.902 17:59:40 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:13:46.902 17:59:40 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@278 -- # remove_spdk_ns 00:13:46.902 17:59:40 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:46.902 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:13:46.902 17:59:40 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:49.437 17:59:42 nvmf_tcp.nvmf_lvs_grow -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:13:49.437 00:13:49.437 real 0m42.283s 00:13:49.437 user 1m6.015s 00:13:49.437 sys 0m9.739s 00:13:49.437 17:59:42 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@1124 -- # xtrace_disable 00:13:49.437 17:59:42 nvmf_tcp.nvmf_lvs_grow -- common/autotest_common.sh@10 -- # set +x 00:13:49.437 ************************************ 00:13:49.437 END TEST nvmf_lvs_grow 00:13:49.437 ************************************ 00:13:49.437 17:59:42 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:13:49.437 17:59:42 nvmf_tcp -- nvmf/nvmf.sh@50 -- # run_test nvmf_bdev_io_wait /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdev_io_wait.sh --transport=tcp 00:13:49.437 17:59:42 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:13:49.437 17:59:42 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:13:49.437 17:59:42 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:13:49.437 ************************************ 00:13:49.437 START TEST nvmf_bdev_io_wait 00:13:49.437 ************************************ 00:13:49.437 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdev_io_wait.sh --transport=tcp 00:13:49.437 * Looking for test storage... 00:13:49.437 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:13:49.437 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:13:49.437 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@7 -- # uname -s 00:13:49.437 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:13:49.437 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:13:49.437 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:13:49.437 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:13:49.437 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- paths/export.sh@5 -- # export PATH 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@47 -- # : 0 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@51 -- # have_pci_nics=0 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@11 -- # MALLOC_BDEV_SIZE=64 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@14 -- # nvmftestinit 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@448 -- # prepare_net_devs 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@410 -- # local -g is_hw=no 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@412 -- # remove_spdk_ns 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@285 -- # xtrace_disable 00:13:49.438 17:59:42 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@291 -- # pci_devs=() 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@291 -- # local -a pci_devs 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@292 -- # pci_net_devs=() 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@293 -- # pci_drivers=() 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@293 -- # local -A pci_drivers 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@295 -- # net_devs=() 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@295 -- # local -ga net_devs 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@296 -- # e810=() 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@296 -- # local -ga e810 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@297 -- # x722=() 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@297 -- # local -ga x722 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@298 -- # mlx=() 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@298 -- # local -ga mlx 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:13:54.712 Found 0000:86:00.0 (0x8086 - 0x159b) 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:13:54.712 Found 0000:86:00.1 (0x8086 - 0x159b) 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@390 -- # [[ up == up ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:13:54.712 Found net devices under 0000:86:00.0: cvl_0_0 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@390 -- # [[ up == up ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:13:54.712 Found net devices under 0000:86:00.1: cvl_0_1 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@414 -- # is_hw=yes 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:13:54.712 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:13:54.713 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:13:54.713 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:13:54.713 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:13:54.713 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:13:54.713 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:13:54.713 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:13:54.713 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:13:54.713 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:13:54.713 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:13:54.713 17:59:47 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:13:54.713 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:13:54.713 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.308 ms 00:13:54.713 00:13:54.713 --- 10.0.0.2 ping statistics --- 00:13:54.713 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:54.713 rtt min/avg/max/mdev = 0.308/0.308/0.308/0.000 ms 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:13:54.713 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:13:54.713 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.105 ms 00:13:54.713 00:13:54.713 --- 10.0.0.1 ping statistics --- 00:13:54.713 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:13:54.713 rtt min/avg/max/mdev = 0.105/0.105/0.105/0.000 ms 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@422 -- # return 0 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@15 -- # nvmfappstart -m 0xF --wait-for-rpc 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@722 -- # xtrace_disable 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@481 -- # nvmfpid=547280 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@482 -- # waitforlisten 547280 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@829 -- # '[' -z 547280 ']' 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@834 -- # local max_retries=100 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:13:54.713 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@838 -- # xtrace_disable 00:13:54.713 17:59:48 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:13:54.713 [2024-07-15 17:59:48.279200] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:13:54.713 [2024-07-15 17:59:48.279262] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:13:54.713 EAL: No free 2048 kB hugepages reported on node 1 00:13:54.713 [2024-07-15 17:59:48.337005] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:13:54.713 [2024-07-15 17:59:48.420258] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:13:54.713 [2024-07-15 17:59:48.420295] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:13:54.713 [2024-07-15 17:59:48.420302] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:13:54.713 [2024-07-15 17:59:48.420308] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:13:54.713 [2024-07-15 17:59:48.420313] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:13:54.713 [2024-07-15 17:59:48.420354] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:13:54.713 [2024-07-15 17:59:48.420450] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:13:54.713 [2024-07-15 17:59:48.420536] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:13:54.713 [2024-07-15 17:59:48.420538] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@862 -- # return 0 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@728 -- # xtrace_disable 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@18 -- # rpc_cmd bdev_set_options -p 5 -c 1 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@559 -- # xtrace_disable 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@19 -- # rpc_cmd framework_start_init 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@559 -- # xtrace_disable 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@20 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@559 -- # xtrace_disable 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:13:55.650 [2024-07-15 17:59:49.206553] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@22 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@559 -- # xtrace_disable 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:13:55.650 Malloc0 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@23 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@559 -- # xtrace_disable 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@559 -- # xtrace_disable 00:13:55.650 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@559 -- # xtrace_disable 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:13:55.651 [2024-07-15 17:59:49.264539] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@28 -- # WRITE_PID=547532 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x10 -i 1 --json /dev/fd/63 -q 128 -o 4096 -w write -t 1 -s 256 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@27 -- # gen_nvmf_target_json 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@30 -- # READ_PID=547534 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@532 -- # config=() 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@532 -- # local subsystem config 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:13:55.651 { 00:13:55.651 "params": { 00:13:55.651 "name": "Nvme$subsystem", 00:13:55.651 "trtype": "$TEST_TRANSPORT", 00:13:55.651 "traddr": "$NVMF_FIRST_TARGET_IP", 00:13:55.651 "adrfam": "ipv4", 00:13:55.651 "trsvcid": "$NVMF_PORT", 00:13:55.651 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:13:55.651 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:13:55.651 "hdgst": ${hdgst:-false}, 00:13:55.651 "ddgst": ${ddgst:-false} 00:13:55.651 }, 00:13:55.651 "method": "bdev_nvme_attach_controller" 00:13:55.651 } 00:13:55.651 EOF 00:13:55.651 )") 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x20 -i 2 --json /dev/fd/63 -q 128 -o 4096 -w read -t 1 -s 256 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@32 -- # FLUSH_PID=547536 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@29 -- # gen_nvmf_target_json 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@532 -- # config=() 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@532 -- # local subsystem config 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:13:55.651 { 00:13:55.651 "params": { 00:13:55.651 "name": "Nvme$subsystem", 00:13:55.651 "trtype": "$TEST_TRANSPORT", 00:13:55.651 "traddr": "$NVMF_FIRST_TARGET_IP", 00:13:55.651 "adrfam": "ipv4", 00:13:55.651 "trsvcid": "$NVMF_PORT", 00:13:55.651 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:13:55.651 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:13:55.651 "hdgst": ${hdgst:-false}, 00:13:55.651 "ddgst": ${ddgst:-false} 00:13:55.651 }, 00:13:55.651 "method": "bdev_nvme_attach_controller" 00:13:55.651 } 00:13:55.651 EOF 00:13:55.651 )") 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x40 -i 3 --json /dev/fd/63 -q 128 -o 4096 -w flush -t 1 -s 256 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@34 -- # UNMAP_PID=547539 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@31 -- # gen_nvmf_target_json 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@35 -- # sync 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@554 -- # cat 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@532 -- # config=() 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@532 -- # local subsystem config 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x80 -i 4 --json /dev/fd/63 -q 128 -o 4096 -w unmap -t 1 -s 256 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@33 -- # gen_nvmf_target_json 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:13:55.651 { 00:13:55.651 "params": { 00:13:55.651 "name": "Nvme$subsystem", 00:13:55.651 "trtype": "$TEST_TRANSPORT", 00:13:55.651 "traddr": "$NVMF_FIRST_TARGET_IP", 00:13:55.651 "adrfam": "ipv4", 00:13:55.651 "trsvcid": "$NVMF_PORT", 00:13:55.651 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:13:55.651 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:13:55.651 "hdgst": ${hdgst:-false}, 00:13:55.651 "ddgst": ${ddgst:-false} 00:13:55.651 }, 00:13:55.651 "method": "bdev_nvme_attach_controller" 00:13:55.651 } 00:13:55.651 EOF 00:13:55.651 )") 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@532 -- # config=() 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@532 -- # local subsystem config 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@554 -- # cat 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:13:55.651 { 00:13:55.651 "params": { 00:13:55.651 "name": "Nvme$subsystem", 00:13:55.651 "trtype": "$TEST_TRANSPORT", 00:13:55.651 "traddr": "$NVMF_FIRST_TARGET_IP", 00:13:55.651 "adrfam": "ipv4", 00:13:55.651 "trsvcid": "$NVMF_PORT", 00:13:55.651 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:13:55.651 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:13:55.651 "hdgst": ${hdgst:-false}, 00:13:55.651 "ddgst": ${ddgst:-false} 00:13:55.651 }, 00:13:55.651 "method": "bdev_nvme_attach_controller" 00:13:55.651 } 00:13:55.651 EOF 00:13:55.651 )") 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@554 -- # cat 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@37 -- # wait 547532 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@554 -- # cat 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@556 -- # jq . 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@556 -- # jq . 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@556 -- # jq . 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@557 -- # IFS=, 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:13:55.651 "params": { 00:13:55.651 "name": "Nvme1", 00:13:55.651 "trtype": "tcp", 00:13:55.651 "traddr": "10.0.0.2", 00:13:55.651 "adrfam": "ipv4", 00:13:55.651 "trsvcid": "4420", 00:13:55.651 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:13:55.651 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:13:55.651 "hdgst": false, 00:13:55.651 "ddgst": false 00:13:55.651 }, 00:13:55.651 "method": "bdev_nvme_attach_controller" 00:13:55.651 }' 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@556 -- # jq . 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@557 -- # IFS=, 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:13:55.651 "params": { 00:13:55.651 "name": "Nvme1", 00:13:55.651 "trtype": "tcp", 00:13:55.651 "traddr": "10.0.0.2", 00:13:55.651 "adrfam": "ipv4", 00:13:55.651 "trsvcid": "4420", 00:13:55.651 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:13:55.651 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:13:55.651 "hdgst": false, 00:13:55.651 "ddgst": false 00:13:55.651 }, 00:13:55.651 "method": "bdev_nvme_attach_controller" 00:13:55.651 }' 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@557 -- # IFS=, 00:13:55.651 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:13:55.651 "params": { 00:13:55.651 "name": "Nvme1", 00:13:55.651 "trtype": "tcp", 00:13:55.651 "traddr": "10.0.0.2", 00:13:55.651 "adrfam": "ipv4", 00:13:55.651 "trsvcid": "4420", 00:13:55.651 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:13:55.652 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:13:55.652 "hdgst": false, 00:13:55.652 "ddgst": false 00:13:55.652 }, 00:13:55.652 "method": "bdev_nvme_attach_controller" 00:13:55.652 }' 00:13:55.652 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@557 -- # IFS=, 00:13:55.652 17:59:49 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:13:55.652 "params": { 00:13:55.652 "name": "Nvme1", 00:13:55.652 "trtype": "tcp", 00:13:55.652 "traddr": "10.0.0.2", 00:13:55.652 "adrfam": "ipv4", 00:13:55.652 "trsvcid": "4420", 00:13:55.652 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:13:55.652 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:13:55.652 "hdgst": false, 00:13:55.652 "ddgst": false 00:13:55.652 }, 00:13:55.652 "method": "bdev_nvme_attach_controller" 00:13:55.652 }' 00:13:55.652 [2024-07-15 17:59:49.315333] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:13:55.652 [2024-07-15 17:59:49.315336] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:13:55.652 [2024-07-15 17:59:49.315334] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:13:55.652 [2024-07-15 17:59:49.315386] [ DPDK EAL parameters: bdevperf -c 0x80 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib[2024-07-15 17:59:49.315387] [ DPDK EAL parameters: bdevperf -c 0x10 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk4 [2024-07-15 17:59:49.315387] [ DPDK EAL parameters: bdevperf -c 0x20 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk1 --proc-type=auto ] 00:13:55.652 .cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk2 --proc-type=auto ] 00:13:55.652 --proc-type=auto ] 00:13:55.652 [2024-07-15 17:59:49.319024] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:13:55.652 [2024-07-15 17:59:49.319071] [ DPDK EAL parameters: bdevperf -c 0x40 -m 256 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk3 --proc-type=auto ] 00:13:55.652 EAL: No free 2048 kB hugepages reported on node 1 00:13:55.910 EAL: No free 2048 kB hugepages reported on node 1 00:13:55.910 [2024-07-15 17:59:49.497617] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:55.910 EAL: No free 2048 kB hugepages reported on node 1 00:13:55.910 [2024-07-15 17:59:49.575483] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 4 00:13:55.910 [2024-07-15 17:59:49.593392] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:56.168 EAL: No free 2048 kB hugepages reported on node 1 00:13:56.168 [2024-07-15 17:59:49.669230] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 5 00:13:56.168 [2024-07-15 17:59:49.690170] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:56.168 [2024-07-15 17:59:49.748191] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:13:56.168 [2024-07-15 17:59:49.780605] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 7 00:13:56.168 [2024-07-15 17:59:49.823989] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 6 00:13:56.428 Running I/O for 1 seconds... 00:13:56.428 Running I/O for 1 seconds... 00:13:56.428 Running I/O for 1 seconds... 00:13:56.428 Running I/O for 1 seconds... 00:13:57.419 00:13:57.419 Latency(us) 00:13:57.419 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:57.419 Job: Nvme1n1 (Core Mask 0x20, workload: read, depth: 128, IO size: 4096) 00:13:57.419 Nvme1n1 : 1.01 13968.37 54.56 0.00 0.00 9133.62 5898.24 17666.23 00:13:57.419 =================================================================================================================== 00:13:57.419 Total : 13968.37 54.56 0.00 0.00 9133.62 5898.24 17666.23 00:13:57.419 00:13:57.419 Latency(us) 00:13:57.419 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:57.419 Job: Nvme1n1 (Core Mask 0x80, workload: unmap, depth: 128, IO size: 4096) 00:13:57.419 Nvme1n1 : 1.01 6131.49 23.95 0.00 0.00 20709.52 10029.86 24846.69 00:13:57.419 =================================================================================================================== 00:13:57.419 Total : 6131.49 23.95 0.00 0.00 20709.52 10029.86 24846.69 00:13:57.419 00:13:57.419 Latency(us) 00:13:57.419 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:57.419 Job: Nvme1n1 (Core Mask 0x40, workload: flush, depth: 128, IO size: 4096) 00:13:57.419 Nvme1n1 : 1.00 245055.55 957.25 0.00 0.00 520.75 213.70 641.11 00:13:57.419 =================================================================================================================== 00:13:57.419 Total : 245055.55 957.25 0.00 0.00 520.75 213.70 641.11 00:13:57.419 00:13:57.419 Latency(us) 00:13:57.419 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:13:57.419 Job: Nvme1n1 (Core Mask 0x10, workload: write, depth: 128, IO size: 4096) 00:13:57.419 Nvme1n1 : 1.00 6281.72 24.54 0.00 0.00 20320.58 4815.47 42398.94 00:13:57.419 =================================================================================================================== 00:13:57.419 Total : 6281.72 24.54 0.00 0.00 20320.58 4815.47 42398.94 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@38 -- # wait 547534 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@39 -- # wait 547536 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@40 -- # wait 547539 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@42 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@559 -- # xtrace_disable 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@44 -- # trap - SIGINT SIGTERM EXIT 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- target/bdev_io_wait.sh@46 -- # nvmftestfini 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@488 -- # nvmfcleanup 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@117 -- # sync 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@120 -- # set +e 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@121 -- # for i in {1..20} 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:13:57.734 rmmod nvme_tcp 00:13:57.734 rmmod nvme_fabrics 00:13:57.734 rmmod nvme_keyring 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@124 -- # set -e 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@125 -- # return 0 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@489 -- # '[' -n 547280 ']' 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@490 -- # killprocess 547280 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@948 -- # '[' -z 547280 ']' 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@952 -- # kill -0 547280 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@953 -- # uname 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 547280 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@966 -- # echo 'killing process with pid 547280' 00:13:57.734 killing process with pid 547280 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@967 -- # kill 547280 00:13:57.734 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@972 -- # wait 547280 00:13:57.991 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:13:57.991 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:13:57.991 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:13:57.991 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:13:57.991 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@278 -- # remove_spdk_ns 00:13:57.991 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:13:57.991 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:13:57.991 17:59:51 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:13:59.891 17:59:53 nvmf_tcp.nvmf_bdev_io_wait -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:13:59.891 00:13:59.891 real 0m10.916s 00:13:59.891 user 0m19.489s 00:13:59.891 sys 0m5.673s 00:13:59.891 17:59:53 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@1124 -- # xtrace_disable 00:13:59.891 17:59:53 nvmf_tcp.nvmf_bdev_io_wait -- common/autotest_common.sh@10 -- # set +x 00:13:59.891 ************************************ 00:13:59.891 END TEST nvmf_bdev_io_wait 00:13:59.891 ************************************ 00:14:00.150 17:59:53 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:14:00.150 17:59:53 nvmf_tcp -- nvmf/nvmf.sh@51 -- # run_test nvmf_queue_depth /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/queue_depth.sh --transport=tcp 00:14:00.150 17:59:53 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:14:00.150 17:59:53 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:14:00.150 17:59:53 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:14:00.150 ************************************ 00:14:00.150 START TEST nvmf_queue_depth 00:14:00.150 ************************************ 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/queue_depth.sh --transport=tcp 00:14:00.150 * Looking for test storage... 00:14:00.150 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- target/queue_depth.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@7 -- # uname -s 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- paths/export.sh@5 -- # export PATH 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@47 -- # : 0 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@51 -- # have_pci_nics=0 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- target/queue_depth.sh@14 -- # MALLOC_BDEV_SIZE=64 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- target/queue_depth.sh@15 -- # MALLOC_BLOCK_SIZE=512 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- target/queue_depth.sh@17 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- target/queue_depth.sh@19 -- # nvmftestinit 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@448 -- # prepare_net_devs 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@410 -- # local -g is_hw=no 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@412 -- # remove_spdk_ns 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@285 -- # xtrace_disable 00:14:00.150 17:59:53 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@291 -- # pci_devs=() 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@291 -- # local -a pci_devs 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@292 -- # pci_net_devs=() 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@293 -- # pci_drivers=() 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@293 -- # local -A pci_drivers 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@295 -- # net_devs=() 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@295 -- # local -ga net_devs 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@296 -- # e810=() 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@296 -- # local -ga e810 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@297 -- # x722=() 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@297 -- # local -ga x722 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@298 -- # mlx=() 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@298 -- # local -ga mlx 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:14:05.422 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:14:05.423 Found 0000:86:00.0 (0x8086 - 0x159b) 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:14:05.423 Found 0000:86:00.1 (0x8086 - 0x159b) 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@390 -- # [[ up == up ]] 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:14:05.423 Found net devices under 0000:86:00.0: cvl_0_0 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@390 -- # [[ up == up ]] 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:14:05.423 Found net devices under 0000:86:00.1: cvl_0_1 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@414 -- # is_hw=yes 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:14:05.423 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:14:05.423 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.250 ms 00:14:05.423 00:14:05.423 --- 10.0.0.2 ping statistics --- 00:14:05.423 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:05.423 rtt min/avg/max/mdev = 0.250/0.250/0.250/0.000 ms 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:14:05.423 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:14:05.423 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.090 ms 00:14:05.423 00:14:05.423 --- 10.0.0.1 ping statistics --- 00:14:05.423 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:05.423 rtt min/avg/max/mdev = 0.090/0.090/0.090/0.000 ms 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@422 -- # return 0 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- target/queue_depth.sh@21 -- # nvmfappstart -m 0x2 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@722 -- # xtrace_disable 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@481 -- # nvmfpid=551316 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@482 -- # waitforlisten 551316 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@829 -- # '[' -z 551316 ']' 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@834 -- # local max_retries=100 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:05.423 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@838 -- # xtrace_disable 00:14:05.423 17:59:58 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:14:05.423 [2024-07-15 17:59:58.762962] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:14:05.423 [2024-07-15 17:59:58.763002] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:05.423 EAL: No free 2048 kB hugepages reported on node 1 00:14:05.423 [2024-07-15 17:59:58.818188] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:05.423 [2024-07-15 17:59:58.894135] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:14:05.423 [2024-07-15 17:59:58.894170] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:14:05.423 [2024-07-15 17:59:58.894177] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:14:05.423 [2024-07-15 17:59:58.894183] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:14:05.423 [2024-07-15 17:59:58.894188] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:14:05.423 [2024-07-15 17:59:58.894207] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@862 -- # return 0 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@728 -- # xtrace_disable 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- target/queue_depth.sh@23 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@559 -- # xtrace_disable 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:14:05.990 [2024-07-15 17:59:59.589845] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- target/queue_depth.sh@24 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@559 -- # xtrace_disable 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:14:05.990 Malloc0 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- target/queue_depth.sh@25 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@559 -- # xtrace_disable 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- target/queue_depth.sh@26 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@559 -- # xtrace_disable 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- target/queue_depth.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@559 -- # xtrace_disable 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:14:05.990 [2024-07-15 17:59:59.648470] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- target/queue_depth.sh@30 -- # bdevperf_pid=551348 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- target/queue_depth.sh@32 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- target/queue_depth.sh@33 -- # waitforlisten 551348 /var/tmp/bdevperf.sock 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@829 -- # '[' -z 551348 ']' 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@834 -- # local max_retries=100 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- target/queue_depth.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 1024 -o 4096 -w verify -t 10 00:14:05.990 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@838 -- # xtrace_disable 00:14:05.990 17:59:59 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:14:05.990 [2024-07-15 17:59:59.697798] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:14:05.990 [2024-07-15 17:59:59.697840] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid551348 ] 00:14:06.249 EAL: No free 2048 kB hugepages reported on node 1 00:14:06.250 [2024-07-15 17:59:59.752801] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:06.250 [2024-07-15 17:59:59.827083] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:14:06.816 18:00:00 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:14:06.816 18:00:00 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@862 -- # return 0 00:14:06.816 18:00:00 nvmf_tcp.nvmf_queue_depth -- target/queue_depth.sh@34 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:14:06.816 18:00:00 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@559 -- # xtrace_disable 00:14:06.816 18:00:00 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:14:07.074 NVMe0n1 00:14:07.074 18:00:00 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:14:07.074 18:00:00 nvmf_tcp.nvmf_queue_depth -- target/queue_depth.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:14:07.074 Running I/O for 10 seconds... 00:14:17.069 00:14:17.069 Latency(us) 00:14:17.070 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:17.070 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 1024, IO size: 4096) 00:14:17.070 Verification LBA range: start 0x0 length 0x4000 00:14:17.070 NVMe0n1 : 10.06 12299.17 48.04 0.00 0.00 82997.84 19831.76 57443.73 00:14:17.070 =================================================================================================================== 00:14:17.070 Total : 12299.17 48.04 0.00 0.00 82997.84 19831.76 57443.73 00:14:17.070 0 00:14:17.335 18:00:10 nvmf_tcp.nvmf_queue_depth -- target/queue_depth.sh@39 -- # killprocess 551348 00:14:17.335 18:00:10 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@948 -- # '[' -z 551348 ']' 00:14:17.335 18:00:10 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@952 -- # kill -0 551348 00:14:17.335 18:00:10 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@953 -- # uname 00:14:17.335 18:00:10 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:14:17.335 18:00:10 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 551348 00:14:17.335 18:00:10 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:14:17.335 18:00:10 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:14:17.335 18:00:10 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@966 -- # echo 'killing process with pid 551348' 00:14:17.335 killing process with pid 551348 00:14:17.335 18:00:10 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@967 -- # kill 551348 00:14:17.335 Received shutdown signal, test time was about 10.000000 seconds 00:14:17.335 00:14:17.335 Latency(us) 00:14:17.335 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:17.335 =================================================================================================================== 00:14:17.335 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:14:17.335 18:00:10 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@972 -- # wait 551348 00:14:17.335 18:00:11 nvmf_tcp.nvmf_queue_depth -- target/queue_depth.sh@41 -- # trap - SIGINT SIGTERM EXIT 00:14:17.335 18:00:11 nvmf_tcp.nvmf_queue_depth -- target/queue_depth.sh@43 -- # nvmftestfini 00:14:17.335 18:00:11 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@488 -- # nvmfcleanup 00:14:17.335 18:00:11 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@117 -- # sync 00:14:17.335 18:00:11 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:14:17.335 18:00:11 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@120 -- # set +e 00:14:17.335 18:00:11 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@121 -- # for i in {1..20} 00:14:17.335 18:00:11 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:14:17.335 rmmod nvme_tcp 00:14:17.335 rmmod nvme_fabrics 00:14:17.601 rmmod nvme_keyring 00:14:17.601 18:00:11 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:14:17.601 18:00:11 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@124 -- # set -e 00:14:17.601 18:00:11 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@125 -- # return 0 00:14:17.601 18:00:11 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@489 -- # '[' -n 551316 ']' 00:14:17.601 18:00:11 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@490 -- # killprocess 551316 00:14:17.601 18:00:11 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@948 -- # '[' -z 551316 ']' 00:14:17.601 18:00:11 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@952 -- # kill -0 551316 00:14:17.601 18:00:11 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@953 -- # uname 00:14:17.601 18:00:11 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:14:17.601 18:00:11 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 551316 00:14:17.601 18:00:11 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:14:17.601 18:00:11 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:14:17.601 18:00:11 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@966 -- # echo 'killing process with pid 551316' 00:14:17.601 killing process with pid 551316 00:14:17.601 18:00:11 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@967 -- # kill 551316 00:14:17.601 18:00:11 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@972 -- # wait 551316 00:14:17.867 18:00:11 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:14:17.867 18:00:11 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:14:17.868 18:00:11 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:14:17.868 18:00:11 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:14:17.868 18:00:11 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@278 -- # remove_spdk_ns 00:14:17.868 18:00:11 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:17.868 18:00:11 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:14:17.868 18:00:11 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:19.812 18:00:13 nvmf_tcp.nvmf_queue_depth -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:14:19.812 00:14:19.812 real 0m19.734s 00:14:19.812 user 0m24.555s 00:14:19.812 sys 0m5.175s 00:14:19.812 18:00:13 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@1124 -- # xtrace_disable 00:14:19.812 18:00:13 nvmf_tcp.nvmf_queue_depth -- common/autotest_common.sh@10 -- # set +x 00:14:19.812 ************************************ 00:14:19.812 END TEST nvmf_queue_depth 00:14:19.812 ************************************ 00:14:19.812 18:00:13 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:14:19.812 18:00:13 nvmf_tcp -- nvmf/nvmf.sh@52 -- # run_test nvmf_target_multipath /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multipath.sh --transport=tcp 00:14:19.812 18:00:13 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:14:19.812 18:00:13 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:14:19.812 18:00:13 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:14:19.812 ************************************ 00:14:19.812 START TEST nvmf_target_multipath 00:14:19.812 ************************************ 00:14:19.812 18:00:13 nvmf_tcp.nvmf_target_multipath -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/multipath.sh --transport=tcp 00:14:20.076 * Looking for test storage... 00:14:20.076 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- target/multipath.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@7 -- # uname -s 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- paths/export.sh@5 -- # export PATH 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@47 -- # : 0 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@51 -- # have_pci_nics=0 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- target/multipath.sh@11 -- # MALLOC_BDEV_SIZE=64 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- target/multipath.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- target/multipath.sh@13 -- # nqn=nqn.2016-06.io.spdk:cnode1 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- target/multipath.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- target/multipath.sh@43 -- # nvmftestinit 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@448 -- # prepare_net_devs 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@410 -- # local -g is_hw=no 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@412 -- # remove_spdk_ns 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@285 -- # xtrace_disable 00:14:20.076 18:00:13 nvmf_tcp.nvmf_target_multipath -- common/autotest_common.sh@10 -- # set +x 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@291 -- # pci_devs=() 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@291 -- # local -a pci_devs 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@292 -- # pci_net_devs=() 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@293 -- # pci_drivers=() 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@293 -- # local -A pci_drivers 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@295 -- # net_devs=() 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@295 -- # local -ga net_devs 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@296 -- # e810=() 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@296 -- # local -ga e810 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@297 -- # x722=() 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@297 -- # local -ga x722 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@298 -- # mlx=() 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@298 -- # local -ga mlx 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:14:25.378 Found 0000:86:00.0 (0x8086 - 0x159b) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:14:25.378 Found 0000:86:00.1 (0x8086 - 0x159b) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@390 -- # [[ up == up ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:14:25.378 Found net devices under 0000:86:00.0: cvl_0_0 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@390 -- # [[ up == up ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:14:25.378 Found net devices under 0000:86:00.1: cvl_0_1 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@414 -- # is_hw=yes 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:14:25.378 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:14:25.379 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:14:25.379 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.199 ms 00:14:25.379 00:14:25.379 --- 10.0.0.2 ping statistics --- 00:14:25.379 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:25.379 rtt min/avg/max/mdev = 0.199/0.199/0.199/0.000 ms 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:14:25.379 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:14:25.379 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.206 ms 00:14:25.379 00:14:25.379 --- 10.0.0.1 ping statistics --- 00:14:25.379 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:25.379 rtt min/avg/max/mdev = 0.206/0.206/0.206/0.000 ms 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@422 -- # return 0 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- target/multipath.sh@45 -- # '[' -z ']' 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- target/multipath.sh@46 -- # echo 'only one NIC for nvmf test' 00:14:25.379 only one NIC for nvmf test 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- target/multipath.sh@47 -- # nvmftestfini 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@488 -- # nvmfcleanup 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@117 -- # sync 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@120 -- # set +e 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@121 -- # for i in {1..20} 00:14:25.379 18:00:18 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:14:25.379 rmmod nvme_tcp 00:14:25.379 rmmod nvme_fabrics 00:14:25.379 rmmod nvme_keyring 00:14:25.379 18:00:19 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:14:25.379 18:00:19 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@124 -- # set -e 00:14:25.379 18:00:19 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@125 -- # return 0 00:14:25.379 18:00:19 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@489 -- # '[' -n '' ']' 00:14:25.379 18:00:19 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:14:25.379 18:00:19 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:14:25.379 18:00:19 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:14:25.379 18:00:19 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:14:25.379 18:00:19 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@278 -- # remove_spdk_ns 00:14:25.379 18:00:19 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:25.379 18:00:19 nvmf_tcp.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:14:25.379 18:00:19 nvmf_tcp.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:27.913 18:00:21 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:14:27.913 18:00:21 nvmf_tcp.nvmf_target_multipath -- target/multipath.sh@48 -- # exit 0 00:14:27.913 18:00:21 nvmf_tcp.nvmf_target_multipath -- target/multipath.sh@1 -- # nvmftestfini 00:14:27.913 18:00:21 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@488 -- # nvmfcleanup 00:14:27.913 18:00:21 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@117 -- # sync 00:14:27.913 18:00:21 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:14:27.913 18:00:21 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@120 -- # set +e 00:14:27.913 18:00:21 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@121 -- # for i in {1..20} 00:14:27.914 18:00:21 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:14:27.914 18:00:21 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:14:27.914 18:00:21 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@124 -- # set -e 00:14:27.914 18:00:21 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@125 -- # return 0 00:14:27.914 18:00:21 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@489 -- # '[' -n '' ']' 00:14:27.914 18:00:21 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:14:27.914 18:00:21 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:14:27.914 18:00:21 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:14:27.914 18:00:21 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:14:27.914 18:00:21 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@278 -- # remove_spdk_ns 00:14:27.914 18:00:21 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:27.914 18:00:21 nvmf_tcp.nvmf_target_multipath -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:14:27.914 18:00:21 nvmf_tcp.nvmf_target_multipath -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:27.914 18:00:21 nvmf_tcp.nvmf_target_multipath -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:14:27.914 00:14:27.914 real 0m7.637s 00:14:27.914 user 0m1.603s 00:14:27.914 sys 0m4.015s 00:14:27.914 18:00:21 nvmf_tcp.nvmf_target_multipath -- common/autotest_common.sh@1124 -- # xtrace_disable 00:14:27.914 18:00:21 nvmf_tcp.nvmf_target_multipath -- common/autotest_common.sh@10 -- # set +x 00:14:27.914 ************************************ 00:14:27.914 END TEST nvmf_target_multipath 00:14:27.914 ************************************ 00:14:27.914 18:00:21 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:14:27.914 18:00:21 nvmf_tcp -- nvmf/nvmf.sh@53 -- # run_test nvmf_zcopy /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh --transport=tcp 00:14:27.914 18:00:21 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:14:27.914 18:00:21 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:14:27.914 18:00:21 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:14:27.914 ************************************ 00:14:27.914 START TEST nvmf_zcopy 00:14:27.914 ************************************ 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh --transport=tcp 00:14:27.914 * Looking for test storage... 00:14:27.914 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@7 -- # uname -s 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- paths/export.sh@5 -- # export PATH 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@47 -- # : 0 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@51 -- # have_pci_nics=0 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@12 -- # nvmftestinit 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@448 -- # prepare_net_devs 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@410 -- # local -g is_hw=no 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@412 -- # remove_spdk_ns 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@285 -- # xtrace_disable 00:14:27.914 18:00:21 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:14:33.226 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:14:33.226 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@291 -- # pci_devs=() 00:14:33.226 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@291 -- # local -a pci_devs 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@292 -- # pci_net_devs=() 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@293 -- # pci_drivers=() 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@293 -- # local -A pci_drivers 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@295 -- # net_devs=() 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@295 -- # local -ga net_devs 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@296 -- # e810=() 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@296 -- # local -ga e810 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@297 -- # x722=() 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@297 -- # local -ga x722 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@298 -- # mlx=() 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@298 -- # local -ga mlx 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:14:33.227 Found 0000:86:00.0 (0x8086 - 0x159b) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:14:33.227 Found 0000:86:00.1 (0x8086 - 0x159b) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@390 -- # [[ up == up ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:14:33.227 Found net devices under 0000:86:00.0: cvl_0_0 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@390 -- # [[ up == up ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:14:33.227 Found net devices under 0000:86:00.1: cvl_0_1 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@414 -- # is_hw=yes 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:14:33.227 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:14:33.227 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.206 ms 00:14:33.227 00:14:33.227 --- 10.0.0.2 ping statistics --- 00:14:33.227 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:33.227 rtt min/avg/max/mdev = 0.206/0.206/0.206/0.000 ms 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:14:33.227 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:14:33.227 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.133 ms 00:14:33.227 00:14:33.227 --- 10.0.0.1 ping statistics --- 00:14:33.227 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:14:33.227 rtt min/avg/max/mdev = 0.133/0.133/0.133/0.000 ms 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@422 -- # return 0 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@13 -- # nvmfappstart -m 0x2 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@722 -- # xtrace_disable 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@481 -- # nvmfpid=560692 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@482 -- # waitforlisten 560692 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@829 -- # '[' -z 560692 ']' 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@834 -- # local max_retries=100 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:14:33.227 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@838 -- # xtrace_disable 00:14:33.227 18:00:26 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:14:33.227 [2024-07-15 18:00:26.571169] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:14:33.227 [2024-07-15 18:00:26.571217] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:14:33.227 EAL: No free 2048 kB hugepages reported on node 1 00:14:33.227 [2024-07-15 18:00:26.630664] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:33.227 [2024-07-15 18:00:26.709357] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:14:33.227 [2024-07-15 18:00:26.709389] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:14:33.228 [2024-07-15 18:00:26.709396] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:14:33.228 [2024-07-15 18:00:26.709402] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:14:33.228 [2024-07-15 18:00:26.709407] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:14:33.228 [2024-07-15 18:00:26.709423] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@862 -- # return 0 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@728 -- # xtrace_disable 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@15 -- # '[' tcp '!=' tcp ']' 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@22 -- # rpc_cmd nvmf_create_transport -t tcp -o -c 0 --zcopy 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@559 -- # xtrace_disable 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:14:33.794 [2024-07-15 18:00:27.415143] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@24 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 10 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@559 -- # xtrace_disable 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@559 -- # xtrace_disable 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:14:33.794 [2024-07-15 18:00:27.435270] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@559 -- # xtrace_disable 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@29 -- # rpc_cmd bdev_malloc_create 32 4096 -b malloc0 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@559 -- # xtrace_disable 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:14:33.794 malloc0 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@30 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@559 -- # xtrace_disable 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -t 10 -q 128 -w verify -o 8192 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@33 -- # gen_nvmf_target_json 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@532 -- # config=() 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@532 -- # local subsystem config 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:14:33.794 { 00:14:33.794 "params": { 00:14:33.794 "name": "Nvme$subsystem", 00:14:33.794 "trtype": "$TEST_TRANSPORT", 00:14:33.794 "traddr": "$NVMF_FIRST_TARGET_IP", 00:14:33.794 "adrfam": "ipv4", 00:14:33.794 "trsvcid": "$NVMF_PORT", 00:14:33.794 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:14:33.794 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:14:33.794 "hdgst": ${hdgst:-false}, 00:14:33.794 "ddgst": ${ddgst:-false} 00:14:33.794 }, 00:14:33.794 "method": "bdev_nvme_attach_controller" 00:14:33.794 } 00:14:33.794 EOF 00:14:33.794 )") 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@554 -- # cat 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@556 -- # jq . 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@557 -- # IFS=, 00:14:33.794 18:00:27 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:14:33.794 "params": { 00:14:33.794 "name": "Nvme1", 00:14:33.794 "trtype": "tcp", 00:14:33.794 "traddr": "10.0.0.2", 00:14:33.794 "adrfam": "ipv4", 00:14:33.794 "trsvcid": "4420", 00:14:33.794 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:14:33.794 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:14:33.794 "hdgst": false, 00:14:33.794 "ddgst": false 00:14:33.794 }, 00:14:33.794 "method": "bdev_nvme_attach_controller" 00:14:33.794 }' 00:14:33.794 [2024-07-15 18:00:27.509991] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:14:33.794 [2024-07-15 18:00:27.510037] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid560747 ] 00:14:34.070 EAL: No free 2048 kB hugepages reported on node 1 00:14:34.070 [2024-07-15 18:00:27.563020] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:34.070 [2024-07-15 18:00:27.637993] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:14:34.329 Running I/O for 10 seconds... 00:14:44.299 00:14:44.299 Latency(us) 00:14:44.299 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:44.299 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 8192) 00:14:44.299 Verification LBA range: start 0x0 length 0x1000 00:14:44.299 Nvme1n1 : 10.01 8659.80 67.65 0.00 0.00 14739.07 1552.92 25188.62 00:14:44.299 =================================================================================================================== 00:14:44.299 Total : 8659.80 67.65 0.00 0.00 14739.07 1552.92 25188.62 00:14:44.558 18:00:38 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@39 -- # perfpid=562569 00:14:44.558 18:00:38 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@41 -- # xtrace_disable 00:14:44.558 18:00:38 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:14:44.558 18:00:38 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/63 -t 5 -q 128 -w randrw -M 50 -o 8192 00:14:44.558 18:00:38 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@37 -- # gen_nvmf_target_json 00:14:44.558 18:00:38 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@532 -- # config=() 00:14:44.558 18:00:38 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@532 -- # local subsystem config 00:14:44.558 18:00:38 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:14:44.558 18:00:38 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:14:44.558 { 00:14:44.558 "params": { 00:14:44.558 "name": "Nvme$subsystem", 00:14:44.558 "trtype": "$TEST_TRANSPORT", 00:14:44.558 "traddr": "$NVMF_FIRST_TARGET_IP", 00:14:44.558 "adrfam": "ipv4", 00:14:44.558 "trsvcid": "$NVMF_PORT", 00:14:44.558 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:14:44.558 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:14:44.558 "hdgst": ${hdgst:-false}, 00:14:44.558 "ddgst": ${ddgst:-false} 00:14:44.558 }, 00:14:44.558 "method": "bdev_nvme_attach_controller" 00:14:44.558 } 00:14:44.558 EOF 00:14:44.558 )") 00:14:44.558 [2024-07-15 18:00:38.195291] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.558 [2024-07-15 18:00:38.195322] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.558 18:00:38 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@554 -- # cat 00:14:44.558 18:00:38 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@556 -- # jq . 00:14:44.558 [2024-07-15 18:00:38.203278] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.558 [2024-07-15 18:00:38.203290] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.558 18:00:38 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@557 -- # IFS=, 00:14:44.558 18:00:38 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:14:44.558 "params": { 00:14:44.558 "name": "Nvme1", 00:14:44.558 "trtype": "tcp", 00:14:44.558 "traddr": "10.0.0.2", 00:14:44.558 "adrfam": "ipv4", 00:14:44.558 "trsvcid": "4420", 00:14:44.558 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:14:44.558 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:14:44.558 "hdgst": false, 00:14:44.558 "ddgst": false 00:14:44.558 }, 00:14:44.558 "method": "bdev_nvme_attach_controller" 00:14:44.558 }' 00:14:44.558 [2024-07-15 18:00:38.211292] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.558 [2024-07-15 18:00:38.211302] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.558 [2024-07-15 18:00:38.219313] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.558 [2024-07-15 18:00:38.219321] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.558 [2024-07-15 18:00:38.227332] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.558 [2024-07-15 18:00:38.227340] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.558 [2024-07-15 18:00:38.233876] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:14:44.559 [2024-07-15 18:00:38.233916] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid562569 ] 00:14:44.559 [2024-07-15 18:00:38.235356] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.559 [2024-07-15 18:00:38.235365] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.559 [2024-07-15 18:00:38.243376] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.559 [2024-07-15 18:00:38.243385] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.559 [2024-07-15 18:00:38.251398] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.559 [2024-07-15 18:00:38.251407] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.559 EAL: No free 2048 kB hugepages reported on node 1 00:14:44.559 [2024-07-15 18:00:38.259422] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.559 [2024-07-15 18:00:38.259432] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.559 [2024-07-15 18:00:38.267443] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.559 [2024-07-15 18:00:38.267454] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.559 [2024-07-15 18:00:38.275465] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.559 [2024-07-15 18:00:38.275477] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.559 [2024-07-15 18:00:38.283486] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.559 [2024-07-15 18:00:38.283496] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.288273] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:14:44.818 [2024-07-15 18:00:38.291506] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.291515] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.299531] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.299544] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.307552] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.307562] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.315572] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.315581] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.323593] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.323602] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.331622] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.331643] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.339637] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.339648] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.347657] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.347667] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.355680] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.355688] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.363702] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.363714] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.364622] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:14:44.818 [2024-07-15 18:00:38.371722] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.371732] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.379758] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.379778] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.387773] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.387785] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.395792] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.395805] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.403812] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.403823] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.411832] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.411843] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.419856] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.419866] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.427879] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.427890] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.435898] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.435908] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.443919] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.443928] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.451952] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.451968] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.459981] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.460000] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.467994] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.468006] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.476025] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.476039] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.484039] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.484053] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.492060] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.492073] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.500080] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.500091] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.508100] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.508109] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 [2024-07-15 18:00:38.516128] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.818 [2024-07-15 18:00:38.516144] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.818 Running I/O for 5 seconds... 00:14:44.818 [2024-07-15 18:00:38.524142] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.819 [2024-07-15 18:00:38.524152] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.819 [2024-07-15 18:00:38.534990] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.819 [2024-07-15 18:00:38.535008] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:44.819 [2024-07-15 18:00:38.544411] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:44.819 [2024-07-15 18:00:38.544429] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.077 [2024-07-15 18:00:38.553887] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.077 [2024-07-15 18:00:38.553906] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.077 [2024-07-15 18:00:38.560758] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.077 [2024-07-15 18:00:38.560775] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.077 [2024-07-15 18:00:38.571858] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.077 [2024-07-15 18:00:38.571875] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.077 [2024-07-15 18:00:38.580725] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.077 [2024-07-15 18:00:38.580743] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.077 [2024-07-15 18:00:38.589273] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.077 [2024-07-15 18:00:38.589290] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.077 [2024-07-15 18:00:38.596261] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.077 [2024-07-15 18:00:38.596278] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.077 [2024-07-15 18:00:38.606453] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.077 [2024-07-15 18:00:38.606471] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.077 [2024-07-15 18:00:38.615735] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.077 [2024-07-15 18:00:38.615753] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.077 [2024-07-15 18:00:38.624440] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.077 [2024-07-15 18:00:38.624461] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.077 [2024-07-15 18:00:38.633783] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.077 [2024-07-15 18:00:38.633801] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.077 [2024-07-15 18:00:38.642102] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.077 [2024-07-15 18:00:38.642120] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.077 [2024-07-15 18:00:38.650738] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.077 [2024-07-15 18:00:38.650754] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.077 [2024-07-15 18:00:38.660007] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.077 [2024-07-15 18:00:38.660024] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.077 [2024-07-15 18:00:38.668844] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.077 [2024-07-15 18:00:38.668861] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.077 [2024-07-15 18:00:38.677984] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.077 [2024-07-15 18:00:38.678002] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.077 [2024-07-15 18:00:38.686711] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.077 [2024-07-15 18:00:38.686728] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.078 [2024-07-15 18:00:38.696038] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.078 [2024-07-15 18:00:38.696056] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.078 [2024-07-15 18:00:38.705303] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.078 [2024-07-15 18:00:38.705320] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.078 [2024-07-15 18:00:38.714467] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.078 [2024-07-15 18:00:38.714484] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.078 [2024-07-15 18:00:38.724268] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.078 [2024-07-15 18:00:38.724286] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.078 [2024-07-15 18:00:38.732737] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.078 [2024-07-15 18:00:38.732754] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.078 [2024-07-15 18:00:38.741892] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.078 [2024-07-15 18:00:38.741909] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.078 [2024-07-15 18:00:38.751171] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.078 [2024-07-15 18:00:38.751189] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.078 [2024-07-15 18:00:38.760705] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.078 [2024-07-15 18:00:38.760722] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.078 [2024-07-15 18:00:38.769928] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.078 [2024-07-15 18:00:38.769946] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.078 [2024-07-15 18:00:38.778597] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.078 [2024-07-15 18:00:38.778615] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.078 [2024-07-15 18:00:38.787435] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.078 [2024-07-15 18:00:38.787454] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.078 [2024-07-15 18:00:38.796181] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.078 [2024-07-15 18:00:38.796199] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.805339] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.805360] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.814011] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.814032] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.823264] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.823285] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.832657] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.832675] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.842138] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.842157] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.850755] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.850774] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.859342] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.859360] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.868630] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.868648] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.877718] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.877737] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.886903] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.886921] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.895709] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.895727] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.904515] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.904534] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.913688] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.913706] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.922400] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.922426] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.931554] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.931572] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.941523] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.941541] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.950440] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.950458] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.959597] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.959615] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.968731] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.968749] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.977878] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.977896] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.987089] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.987107] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:38.995659] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:38.995678] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:39.004897] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:39.004915] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:39.014138] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:39.014157] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:39.022821] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:39.022839] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:39.032038] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:39.032056] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:39.041892] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:39.041910] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:39.050462] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:39.050480] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.337 [2024-07-15 18:00:39.059601] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.337 [2024-07-15 18:00:39.059619] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.068128] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.068147] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.076934] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.076953] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.085535] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.085553] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.092707] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.092725] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.103127] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.103145] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.111785] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.111803] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.121201] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.121220] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.130014] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.130032] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.138712] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.138730] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.147920] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.147937] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.157247] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.157265] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.165935] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.165953] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.174572] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.174589] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.183243] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.183261] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.192544] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.192561] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.201871] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.201888] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.211077] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.211095] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.219774] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.219791] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.228570] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.228588] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.237346] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.237364] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.246626] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.246644] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.255848] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.255866] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.264604] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.264621] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.273871] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.273889] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.283275] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.283296] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.292516] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.292536] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.301158] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.301175] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.309967] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.309985] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.596 [2024-07-15 18:00:39.317352] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.596 [2024-07-15 18:00:39.317369] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.327151] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.327170] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.335709] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.335726] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.344231] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.344249] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.352799] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.352817] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.361911] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.361928] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.370861] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.370879] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.379573] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.379590] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.389065] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.389083] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.397804] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.397822] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.406807] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.406825] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.416196] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.416213] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.424814] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.424832] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.434095] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.434112] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.443303] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.443321] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.451891] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.451908] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.461251] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.461269] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.470550] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.470571] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.479859] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.479878] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.488540] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.488557] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.495536] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.495552] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.506518] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.506535] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.515420] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.515440] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.524441] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.524459] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.533037] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.533055] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.541659] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.541677] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.548527] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.548544] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.559629] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.559647] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.568460] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.568479] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:45.856 [2024-07-15 18:00:39.577605] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:45.856 [2024-07-15 18:00:39.577624] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.586688] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.586707] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.595119] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.595137] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.603646] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.603664] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.612185] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.612204] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.621432] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.621450] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.630931] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.630949] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.637918] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.637939] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.648998] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.649016] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.657956] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.657973] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.666780] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.666798] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.675326] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.675344] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.684580] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.684597] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.693742] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.693760] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.702339] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.702356] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.711615] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.711632] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.721003] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.721021] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.729843] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.729861] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.739070] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.739088] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.748378] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.748395] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.757647] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.757665] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.766340] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.766358] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.775681] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.775698] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.785631] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.785648] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.794378] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.794396] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.803585] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.803603] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.813069] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.813091] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.822304] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.822322] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.115 [2024-07-15 18:00:39.829431] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.115 [2024-07-15 18:00:39.829449] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.116 [2024-07-15 18:00:39.839858] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.116 [2024-07-15 18:00:39.839876] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:39.848694] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:39.848712] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:39.857539] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:39.857556] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:39.867114] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:39.867132] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:39.874306] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:39.874323] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:39.884802] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:39.884820] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:39.893901] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:39.893919] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:39.903407] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:39.903425] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:39.912061] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:39.912078] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:39.920690] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:39.920708] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:39.929926] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:39.929943] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:39.938603] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:39.938621] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:39.947254] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:39.947271] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:39.955757] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:39.955774] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:39.965038] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:39.965056] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:39.974393] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:39.974410] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:39.983582] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:39.983603] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:39.992796] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:39.992814] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:40.001985] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:40.002002] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:40.011172] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:40.011189] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:40.020239] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:40.020257] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:40.027093] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:40.027110] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:40.038380] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:40.038399] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:40.047353] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:40.047370] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:40.055926] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:40.055944] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:40.064839] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:40.064859] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:40.073393] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:40.073419] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:40.083505] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:40.083525] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.374 [2024-07-15 18:00:40.092942] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.374 [2024-07-15 18:00:40.092960] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.102270] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.102289] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.109189] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.109206] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.119289] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.119306] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.128571] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.128588] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.137047] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.137065] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.146172] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.146190] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.155298] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.155316] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.164598] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.164615] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.173242] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.173260] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.181949] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.181967] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.191271] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.191290] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.200702] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.200722] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.209388] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.209407] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.218617] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.218636] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.227353] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.227371] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.234770] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.234787] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.245235] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.245254] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.254167] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.254187] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.262849] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.262867] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.271485] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.271503] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.280238] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.280256] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.288935] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.288953] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.298421] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.298440] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.307662] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.307681] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.317150] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.317169] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.326531] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.326549] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.335822] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.335840] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.345142] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.345161] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.633 [2024-07-15 18:00:40.353754] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.633 [2024-07-15 18:00:40.353771] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.362365] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.362384] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.371040] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.371058] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.378010] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.378027] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.388991] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.389010] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.398578] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.398596] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.407275] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.407293] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.415898] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.415915] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.424630] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.424648] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.434060] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.434079] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.442775] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.442792] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.452093] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.452111] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.461558] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.461577] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.470933] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.470951] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.479591] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.479608] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.488414] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.488431] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.497615] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.497633] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.507131] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.507149] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.516412] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.516430] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.525533] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.525555] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.535522] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.535540] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.544283] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.544301] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.551550] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.551568] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.561363] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.561381] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.570218] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.891 [2024-07-15 18:00:40.570242] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.891 [2024-07-15 18:00:40.578463] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.892 [2024-07-15 18:00:40.578480] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.892 [2024-07-15 18:00:40.587878] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.892 [2024-07-15 18:00:40.587895] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.892 [2024-07-15 18:00:40.596535] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.892 [2024-07-15 18:00:40.596554] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.892 [2024-07-15 18:00:40.605830] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.892 [2024-07-15 18:00:40.605848] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:46.892 [2024-07-15 18:00:40.615198] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:46.892 [2024-07-15 18:00:40.615215] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.165 [2024-07-15 18:00:40.623929] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.165 [2024-07-15 18:00:40.623948] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.165 [2024-07-15 18:00:40.633371] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.165 [2024-07-15 18:00:40.633389] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.165 [2024-07-15 18:00:40.642002] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.165 [2024-07-15 18:00:40.642019] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.165 [2024-07-15 18:00:40.651458] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.165 [2024-07-15 18:00:40.651476] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.165 [2024-07-15 18:00:40.660192] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.165 [2024-07-15 18:00:40.660211] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.165 [2024-07-15 18:00:40.668946] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.165 [2024-07-15 18:00:40.668963] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.165 [2024-07-15 18:00:40.678049] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.165 [2024-07-15 18:00:40.678068] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.165 [2024-07-15 18:00:40.686585] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.165 [2024-07-15 18:00:40.686602] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.165 [2024-07-15 18:00:40.695823] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.165 [2024-07-15 18:00:40.695841] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.166 [2024-07-15 18:00:40.704414] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.166 [2024-07-15 18:00:40.704431] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.166 [2024-07-15 18:00:40.713627] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.166 [2024-07-15 18:00:40.713644] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.166 [2024-07-15 18:00:40.722362] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.166 [2024-07-15 18:00:40.722380] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.166 [2024-07-15 18:00:40.731631] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.166 [2024-07-15 18:00:40.731652] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.166 [2024-07-15 18:00:40.740862] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.166 [2024-07-15 18:00:40.740880] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.166 [2024-07-15 18:00:40.749959] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.166 [2024-07-15 18:00:40.749977] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.166 [2024-07-15 18:00:40.758584] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.166 [2024-07-15 18:00:40.758601] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.166 [2024-07-15 18:00:40.768187] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.166 [2024-07-15 18:00:40.768205] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.166 [2024-07-15 18:00:40.776897] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.166 [2024-07-15 18:00:40.776915] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.166 [2024-07-15 18:00:40.785721] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.166 [2024-07-15 18:00:40.785739] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.166 [2024-07-15 18:00:40.794784] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.166 [2024-07-15 18:00:40.794801] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.166 [2024-07-15 18:00:40.803970] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.166 [2024-07-15 18:00:40.803988] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.167 [2024-07-15 18:00:40.813181] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.167 [2024-07-15 18:00:40.813198] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.167 [2024-07-15 18:00:40.822094] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.167 [2024-07-15 18:00:40.822112] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.167 [2024-07-15 18:00:40.830635] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.167 [2024-07-15 18:00:40.830663] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.167 [2024-07-15 18:00:40.839344] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.167 [2024-07-15 18:00:40.839361] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.167 [2024-07-15 18:00:40.848151] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.167 [2024-07-15 18:00:40.848168] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.167 [2024-07-15 18:00:40.856832] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.167 [2024-07-15 18:00:40.856849] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.167 [2024-07-15 18:00:40.865533] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.167 [2024-07-15 18:00:40.865551] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.167 [2024-07-15 18:00:40.874371] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.167 [2024-07-15 18:00:40.874388] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.167 [2024-07-15 18:00:40.882991] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.167 [2024-07-15 18:00:40.883009] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.434 [2024-07-15 18:00:40.892208] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.434 [2024-07-15 18:00:40.892231] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:40.901241] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:40.901259] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:40.910409] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:40.910426] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:40.919115] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:40.919133] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:40.927974] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:40.927991] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:40.936485] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:40.936502] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:40.945086] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:40.945103] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:40.953709] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:40.953726] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:40.962861] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:40.962878] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:40.969786] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:40.969803] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:40.980136] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:40.980153] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:40.988829] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:40.988846] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:40.995660] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:40.995681] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:41.006495] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:41.006512] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:41.015405] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:41.015422] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:41.024871] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:41.024889] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:41.034210] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:41.034236] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:41.042780] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:41.042797] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:41.052130] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:41.052148] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:41.061353] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:41.061371] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:41.070660] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:41.070677] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:41.079902] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:41.079919] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:41.088404] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:41.088422] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:41.097794] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:41.097812] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:41.107241] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:41.107258] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:41.116156] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:41.116173] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:41.125456] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:41.125473] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:41.134877] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:41.134895] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:41.143451] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:41.143468] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:41.152159] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:41.152176] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.435 [2024-07-15 18:00:41.161374] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.435 [2024-07-15 18:00:41.161392] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.170605] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.170628] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.179952] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.179969] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.188562] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.188579] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.197247] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.197264] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.206057] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.206075] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.212979] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.212996] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.223282] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.223301] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.232691] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.232709] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.241970] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.241987] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.251054] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.251071] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.260339] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.260358] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.268946] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.268963] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.278172] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.278190] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.286939] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.286956] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.296033] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.296051] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.302921] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.302938] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.313208] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.313232] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.322537] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.322558] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.331760] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.331778] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.341236] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.341259] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.695 [2024-07-15 18:00:41.350439] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.695 [2024-07-15 18:00:41.350457] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.696 [2024-07-15 18:00:41.358960] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.696 [2024-07-15 18:00:41.358977] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.696 [2024-07-15 18:00:41.367666] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.696 [2024-07-15 18:00:41.367683] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.696 [2024-07-15 18:00:41.376404] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.696 [2024-07-15 18:00:41.376422] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.696 [2024-07-15 18:00:41.385997] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.696 [2024-07-15 18:00:41.386015] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.696 [2024-07-15 18:00:41.395335] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.696 [2024-07-15 18:00:41.395353] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.696 [2024-07-15 18:00:41.402456] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.696 [2024-07-15 18:00:41.402473] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.696 [2024-07-15 18:00:41.413607] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.696 [2024-07-15 18:00:41.413624] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.696 [2024-07-15 18:00:41.422497] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.696 [2024-07-15 18:00:41.422515] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.430958] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.430976] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.440164] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.440181] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.448827] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.448846] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.458152] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.458170] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.467568] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.467585] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.476976] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.476993] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.486049] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.486066] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.494853] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.494871] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.504236] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.504270] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.513719] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.513740] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.522466] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.522483] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.531057] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.531074] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.540262] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.540279] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.549442] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.549460] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.558731] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.558748] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.565622] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.565639] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.575687] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.575705] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.584981] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.585001] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.593474] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.593492] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.602188] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.602207] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.610748] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.610766] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.620036] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.620055] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.629245] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.629264] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.638631] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.638649] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.647302] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.647320] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.655989] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.656007] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.664835] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.664854] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.673368] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.673386] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:47.955 [2024-07-15 18:00:41.682057] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:47.955 [2024-07-15 18:00:41.682076] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.215 [2024-07-15 18:00:41.691382] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.215 [2024-07-15 18:00:41.691401] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.215 [2024-07-15 18:00:41.700595] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.215 [2024-07-15 18:00:41.700613] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.215 [2024-07-15 18:00:41.709722] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.215 [2024-07-15 18:00:41.709741] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.215 [2024-07-15 18:00:41.718491] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.215 [2024-07-15 18:00:41.718509] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.215 [2024-07-15 18:00:41.727891] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.215 [2024-07-15 18:00:41.727909] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.215 [2024-07-15 18:00:41.736616] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.215 [2024-07-15 18:00:41.736635] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.215 [2024-07-15 18:00:41.746080] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.215 [2024-07-15 18:00:41.746099] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.215 [2024-07-15 18:00:41.755457] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.215 [2024-07-15 18:00:41.755476] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.215 [2024-07-15 18:00:41.764645] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.215 [2024-07-15 18:00:41.764664] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.215 [2024-07-15 18:00:41.773270] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.215 [2024-07-15 18:00:41.773288] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.215 [2024-07-15 18:00:41.782002] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.215 [2024-07-15 18:00:41.782021] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.215 [2024-07-15 18:00:41.790549] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.215 [2024-07-15 18:00:41.790567] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.215 [2024-07-15 18:00:41.799912] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.215 [2024-07-15 18:00:41.799930] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.215 [2024-07-15 18:00:41.808896] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.215 [2024-07-15 18:00:41.808915] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.216 [2024-07-15 18:00:41.818370] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.216 [2024-07-15 18:00:41.818388] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.216 [2024-07-15 18:00:41.827991] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.216 [2024-07-15 18:00:41.828009] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.216 [2024-07-15 18:00:41.837075] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.216 [2024-07-15 18:00:41.837093] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.216 [2024-07-15 18:00:41.846343] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.216 [2024-07-15 18:00:41.846361] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.216 [2024-07-15 18:00:41.855364] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.216 [2024-07-15 18:00:41.855383] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.216 [2024-07-15 18:00:41.864568] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.216 [2024-07-15 18:00:41.864586] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.216 [2024-07-15 18:00:41.873174] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.216 [2024-07-15 18:00:41.873192] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.216 [2024-07-15 18:00:41.882463] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.216 [2024-07-15 18:00:41.882481] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.216 [2024-07-15 18:00:41.891743] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.216 [2024-07-15 18:00:41.891761] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.216 [2024-07-15 18:00:41.900761] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.216 [2024-07-15 18:00:41.900779] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.216 [2024-07-15 18:00:41.909924] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.216 [2024-07-15 18:00:41.909942] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.216 [2024-07-15 18:00:41.918566] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.216 [2024-07-15 18:00:41.918584] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.216 [2024-07-15 18:00:41.927211] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.216 [2024-07-15 18:00:41.927236] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.216 [2024-07-15 18:00:41.935660] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.216 [2024-07-15 18:00:41.935679] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.475 [2024-07-15 18:00:41.945116] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.475 [2024-07-15 18:00:41.945135] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.475 [2024-07-15 18:00:41.954840] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.475 [2024-07-15 18:00:41.954858] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.475 [2024-07-15 18:00:41.963667] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.475 [2024-07-15 18:00:41.963684] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.475 [2024-07-15 18:00:41.973155] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.475 [2024-07-15 18:00:41.973174] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.475 [2024-07-15 18:00:41.982405] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.475 [2024-07-15 18:00:41.982423] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.475 [2024-07-15 18:00:41.990988] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.475 [2024-07-15 18:00:41.991006] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.475 [2024-07-15 18:00:41.999526] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.475 [2024-07-15 18:00:41.999544] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.475 [2024-07-15 18:00:42.008785] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.475 [2024-07-15 18:00:42.008803] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.475 [2024-07-15 18:00:42.018205] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.475 [2024-07-15 18:00:42.018223] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.475 [2024-07-15 18:00:42.027400] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.475 [2024-07-15 18:00:42.027418] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.475 [2024-07-15 18:00:42.036507] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.475 [2024-07-15 18:00:42.036524] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.475 [2024-07-15 18:00:42.046383] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.475 [2024-07-15 18:00:42.046400] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.475 [2024-07-15 18:00:42.055038] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.475 [2024-07-15 18:00:42.055055] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.475 [2024-07-15 18:00:42.064231] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.475 [2024-07-15 18:00:42.064249] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.475 [2024-07-15 18:00:42.073000] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.475 [2024-07-15 18:00:42.073018] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.475 [2024-07-15 18:00:42.081576] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.475 [2024-07-15 18:00:42.081593] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.475 [2024-07-15 18:00:42.090181] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.475 [2024-07-15 18:00:42.090198] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.475 [2024-07-15 18:00:42.098639] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.475 [2024-07-15 18:00:42.098655] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.475 [2024-07-15 18:00:42.107921] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.475 [2024-07-15 18:00:42.107938] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.475 [2024-07-15 18:00:42.117096] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.476 [2024-07-15 18:00:42.117114] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.476 [2024-07-15 18:00:42.126414] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.476 [2024-07-15 18:00:42.126431] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.476 [2024-07-15 18:00:42.134968] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.476 [2024-07-15 18:00:42.134985] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.476 [2024-07-15 18:00:42.144020] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.476 [2024-07-15 18:00:42.144036] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.476 [2024-07-15 18:00:42.152620] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.476 [2024-07-15 18:00:42.152637] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.476 [2024-07-15 18:00:42.161088] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.476 [2024-07-15 18:00:42.161104] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.476 [2024-07-15 18:00:42.170281] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.476 [2024-07-15 18:00:42.170298] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.476 [2024-07-15 18:00:42.178966] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.476 [2024-07-15 18:00:42.178984] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.476 [2024-07-15 18:00:42.188220] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.476 [2024-07-15 18:00:42.188246] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.476 [2024-07-15 18:00:42.197246] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.476 [2024-07-15 18:00:42.197279] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.207082] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.207101] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.221238] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.221272] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.230000] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.230017] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.238799] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.238817] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.248218] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.248240] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.257569] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.257587] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.266334] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.266351] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.274990] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.275008] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.284243] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.284261] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.293607] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.293624] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.302795] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.302812] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.311479] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.311497] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.320604] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.320621] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.329208] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.329232] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.338271] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.338290] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.347417] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.347437] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.356875] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.356893] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.365547] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.365569] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.374727] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.374745] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.384072] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.384090] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.393294] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.393312] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.402712] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.402729] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.411865] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.411882] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.421252] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.421269] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.430629] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.430646] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.439727] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.439744] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.449042] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.449059] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.735 [2024-07-15 18:00:42.457645] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.735 [2024-07-15 18:00:42.457663] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.466907] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.466925] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.476277] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.476294] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.485431] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.485449] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.494110] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.494128] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.502875] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.502892] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.511620] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.511638] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.520413] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.520431] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.529827] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.529845] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.538444] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.538466] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.547803] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.547821] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.557452] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.557469] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.566220] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.566244] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.575442] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.575459] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.584586] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.584603] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.593196] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.593213] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.601884] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.601902] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.611209] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.611231] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.619877] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.619895] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.629383] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.629401] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.638677] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.638695] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.648015] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.648032] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.657538] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.657556] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.667128] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.667146] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.676445] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.676463] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.685717] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.685735] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.994 [2024-07-15 18:00:42.694911] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.994 [2024-07-15 18:00:42.694929] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.995 [2024-07-15 18:00:42.703875] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.995 [2024-07-15 18:00:42.703892] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.995 [2024-07-15 18:00:42.712529] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.995 [2024-07-15 18:00:42.712550] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:48.995 [2024-07-15 18:00:42.720724] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:48.995 [2024-07-15 18:00:42.720741] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.729691] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.729710] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.736471] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.736488] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.747560] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.747578] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.756071] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.756089] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.765450] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.765468] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.774519] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.774537] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.783143] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.783161] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.791765] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.791783] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.801066] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.801084] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.810317] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.810335] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.819132] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.819149] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.827768] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.827786] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.836969] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.836987] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.846213] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.846236] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.854862] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.854880] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.864060] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.864078] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.872769] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.872786] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.881987] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.882009] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.891325] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.891343] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.899815] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.899833] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.908368] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.908385] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.916867] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.916884] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.926257] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.926274] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.936131] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.936148] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.944993] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.945010] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.954266] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.954284] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.963690] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.963708] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.253 [2024-07-15 18:00:42.972895] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.253 [2024-07-15 18:00:42.972913] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.512 [2024-07-15 18:00:42.981531] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.512 [2024-07-15 18:00:42.981551] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.512 [2024-07-15 18:00:42.990245] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.512 [2024-07-15 18:00:42.990264] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.512 [2024-07-15 18:00:42.998753] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.512 [2024-07-15 18:00:42.998771] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.512 [2024-07-15 18:00:43.008148] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.512 [2024-07-15 18:00:43.008166] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.512 [2024-07-15 18:00:43.016698] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.512 [2024-07-15 18:00:43.016717] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.512 [2024-07-15 18:00:43.025459] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.512 [2024-07-15 18:00:43.025478] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.034589] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.034608] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.043931] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.043949] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.053265] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.053284] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.062514] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.062532] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.071258] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.071276] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.079918] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.079936] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.088458] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.088476] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.097910] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.097928] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.106554] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.106571] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.115236] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.115254] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.123874] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.123892] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.133111] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.133128] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.142589] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.142607] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.151331] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.151349] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.159778] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.159795] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.166585] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.166603] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.177765] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.177783] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.186685] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.186702] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.196432] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.196450] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.204752] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.204769] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.213393] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.213411] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.222152] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.222171] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.231474] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.231492] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.513 [2024-07-15 18:00:43.240059] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.513 [2024-07-15 18:00:43.240077] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.248655] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.248674] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.257554] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.257573] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.266193] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.266212] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.275732] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.275751] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.284291] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.284309] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.293052] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.293070] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.302389] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.302407] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.311676] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.311693] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.320863] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.320881] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.330126] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.330144] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.339516] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.339534] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.348705] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.348723] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.357373] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.357392] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.366535] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.366553] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.375202] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.375220] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.383783] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.383802] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.392324] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.392342] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.399251] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.399268] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.409868] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.409886] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.418703] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.418720] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.428018] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.428035] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.437176] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.437194] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.446510] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.446528] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.455687] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.455704] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.465493] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.465511] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.474284] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.474302] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.483631] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.483651] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:49.771 [2024-07-15 18:00:43.493018] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:49.771 [2024-07-15 18:00:43.493037] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.502451] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.502469] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.511093] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.511110] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.520424] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.520442] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.529545] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.529562] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.538514] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.538532] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.544526] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.544543] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 00:14:50.031 Latency(us) 00:14:50.031 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:14:50.031 Job: Nvme1n1 (Core Mask 0x1, workload: randrw, percentage: 50, depth: 128, IO size: 8192) 00:14:50.031 Nvme1n1 : 5.01 16579.49 129.53 0.00 0.00 7712.79 3305.29 17324.30 00:14:50.031 =================================================================================================================== 00:14:50.031 Total : 16579.49 129.53 0.00 0.00 7712.79 3305.29 17324.30 00:14:50.031 [2024-07-15 18:00:43.552548] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.552561] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.560566] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.560577] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.568595] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.568609] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.576618] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.576634] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.584636] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.584650] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.592657] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.592670] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.600674] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.600685] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.608695] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.608707] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.616719] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.616731] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.624739] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.624751] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.632761] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.632773] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.640782] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.640793] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.648802] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.648812] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.656823] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.656832] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.664847] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.664860] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.672868] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.672879] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.680888] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.680906] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.031 [2024-07-15 18:00:43.688906] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.031 [2024-07-15 18:00:43.688915] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.032 [2024-07-15 18:00:43.696929] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.032 [2024-07-15 18:00:43.696938] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.032 [2024-07-15 18:00:43.704951] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.032 [2024-07-15 18:00:43.704962] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.032 [2024-07-15 18:00:43.712972] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.032 [2024-07-15 18:00:43.712981] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.032 [2024-07-15 18:00:43.720993] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.032 [2024-07-15 18:00:43.721001] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.032 [2024-07-15 18:00:43.729015] subsystem.c:2054:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Requested NSID 1 already in use 00:14:50.032 [2024-07-15 18:00:43.729025] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:14:50.032 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/zcopy.sh: line 42: kill: (562569) - No such process 00:14:50.032 18:00:43 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@49 -- # wait 562569 00:14:50.032 18:00:43 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@52 -- # rpc_cmd nvmf_subsystem_remove_ns nqn.2016-06.io.spdk:cnode1 1 00:14:50.032 18:00:43 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@559 -- # xtrace_disable 00:14:50.032 18:00:43 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:14:50.032 18:00:43 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:14:50.032 18:00:43 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@53 -- # rpc_cmd bdev_delay_create -b malloc0 -d delay0 -r 1000000 -t 1000000 -w 1000000 -n 1000000 00:14:50.032 18:00:43 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@559 -- # xtrace_disable 00:14:50.032 18:00:43 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:14:50.032 delay0 00:14:50.032 18:00:43 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:14:50.032 18:00:43 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@54 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 delay0 -n 1 00:14:50.032 18:00:43 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@559 -- # xtrace_disable 00:14:50.032 18:00:43 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:14:50.032 18:00:43 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:14:50.032 18:00:43 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -c 0x1 -t 5 -q 64 -w randrw -M 50 -l warning -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 ns:1' 00:14:50.291 EAL: No free 2048 kB hugepages reported on node 1 00:14:50.291 [2024-07-15 18:00:43.850667] nvme_fabric.c: 295:nvme_fabric_discover_probe: *WARNING*: Skipping unsupported current discovery service or discovery service referral 00:14:58.415 Initializing NVMe Controllers 00:14:58.415 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:14:58.415 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:14:58.415 Initialization complete. Launching workers. 00:14:58.415 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 I/O completed: 320, failed: 5721 00:14:58.415 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) abort submitted 6004, failed to submit 37 00:14:58.415 success 5848, unsuccess 156, failed 0 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@59 -- # trap - SIGINT SIGTERM EXIT 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- target/zcopy.sh@60 -- # nvmftestfini 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@488 -- # nvmfcleanup 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@117 -- # sync 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@120 -- # set +e 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@121 -- # for i in {1..20} 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:14:58.415 rmmod nvme_tcp 00:14:58.415 rmmod nvme_fabrics 00:14:58.415 rmmod nvme_keyring 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@124 -- # set -e 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@125 -- # return 0 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@489 -- # '[' -n 560692 ']' 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@490 -- # killprocess 560692 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@948 -- # '[' -z 560692 ']' 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@952 -- # kill -0 560692 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@953 -- # uname 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 560692 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@966 -- # echo 'killing process with pid 560692' 00:14:58.415 killing process with pid 560692 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@967 -- # kill 560692 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@972 -- # wait 560692 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@278 -- # remove_spdk_ns 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:14:58.415 18:00:50 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:59.348 18:00:53 nvmf_tcp.nvmf_zcopy -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:14:59.348 00:14:59.348 real 0m31.845s 00:14:59.348 user 0m43.243s 00:14:59.348 sys 0m10.824s 00:14:59.348 18:00:53 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@1124 -- # xtrace_disable 00:14:59.348 18:00:53 nvmf_tcp.nvmf_zcopy -- common/autotest_common.sh@10 -- # set +x 00:14:59.348 ************************************ 00:14:59.348 END TEST nvmf_zcopy 00:14:59.348 ************************************ 00:14:59.348 18:00:53 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:14:59.348 18:00:53 nvmf_tcp -- nvmf/nvmf.sh@54 -- # run_test nvmf_nmic /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nmic.sh --transport=tcp 00:14:59.348 18:00:53 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:14:59.348 18:00:53 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:14:59.348 18:00:53 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:14:59.606 ************************************ 00:14:59.606 START TEST nvmf_nmic 00:14:59.606 ************************************ 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/nmic.sh --transport=tcp 00:14:59.606 * Looking for test storage... 00:14:59.606 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- target/nmic.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@7 -- # uname -s 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:59.606 18:00:53 nvmf_tcp.nvmf_nmic -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- paths/export.sh@5 -- # export PATH 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@47 -- # : 0 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@51 -- # have_pci_nics=0 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- target/nmic.sh@11 -- # MALLOC_BDEV_SIZE=64 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- target/nmic.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- target/nmic.sh@14 -- # nvmftestinit 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@448 -- # prepare_net_devs 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@410 -- # local -g is_hw=no 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@412 -- # remove_spdk_ns 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@285 -- # xtrace_disable 00:14:59.607 18:00:53 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@291 -- # pci_devs=() 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@291 -- # local -a pci_devs 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@292 -- # pci_net_devs=() 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@293 -- # pci_drivers=() 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@293 -- # local -A pci_drivers 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@295 -- # net_devs=() 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@295 -- # local -ga net_devs 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@296 -- # e810=() 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@296 -- # local -ga e810 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@297 -- # x722=() 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@297 -- # local -ga x722 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@298 -- # mlx=() 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@298 -- # local -ga mlx 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:15:04.916 Found 0000:86:00.0 (0x8086 - 0x159b) 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:15:04.916 Found 0000:86:00.1 (0x8086 - 0x159b) 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@390 -- # [[ up == up ]] 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:15:04.916 Found net devices under 0000:86:00.0: cvl_0_0 00:15:04.916 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@390 -- # [[ up == up ]] 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:15:04.917 Found net devices under 0000:86:00.1: cvl_0_1 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@414 -- # is_hw=yes 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:15:04.917 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:15:04.917 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.180 ms 00:15:04.917 00:15:04.917 --- 10.0.0.2 ping statistics --- 00:15:04.917 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:04.917 rtt min/avg/max/mdev = 0.180/0.180/0.180/0.000 ms 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:15:04.917 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:15:04.917 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.168 ms 00:15:04.917 00:15:04.917 --- 10.0.0.1 ping statistics --- 00:15:04.917 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:04.917 rtt min/avg/max/mdev = 0.168/0.168/0.168/0.000 ms 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@422 -- # return 0 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- target/nmic.sh@15 -- # nvmfappstart -m 0xF 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@722 -- # xtrace_disable 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@481 -- # nvmfpid=568004 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@482 -- # waitforlisten 568004 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@829 -- # '[' -z 568004 ']' 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@834 -- # local max_retries=100 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:04.917 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@838 -- # xtrace_disable 00:15:04.917 18:00:58 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:15:04.917 [2024-07-15 18:00:58.571759] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:15:04.917 [2024-07-15 18:00:58.571806] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:04.917 EAL: No free 2048 kB hugepages reported on node 1 00:15:04.917 [2024-07-15 18:00:58.628372] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:15:05.175 [2024-07-15 18:00:58.710215] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:15:05.175 [2024-07-15 18:00:58.710253] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:15:05.175 [2024-07-15 18:00:58.710260] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:15:05.175 [2024-07-15 18:00:58.710265] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:15:05.175 [2024-07-15 18:00:58.710271] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:15:05.175 [2024-07-15 18:00:58.710304] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:15:05.175 [2024-07-15 18:00:58.710400] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:15:05.175 [2024-07-15 18:00:58.710492] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:15:05.175 [2024-07-15 18:00:58.710493] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@862 -- # return 0 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@728 -- # xtrace_disable 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- target/nmic.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:15:05.742 [2024-07-15 18:00:59.426267] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- target/nmic.sh@20 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:15:05.742 Malloc0 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- target/nmic.sh@21 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- target/nmic.sh@22 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:05.742 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:15:06.001 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:06.001 18:00:59 nvmf_tcp.nvmf_nmic -- target/nmic.sh@23 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:15:06.001 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:06.001 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:15:06.001 [2024-07-15 18:00:59.478107] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:15:06.001 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:06.001 18:00:59 nvmf_tcp.nvmf_nmic -- target/nmic.sh@25 -- # echo 'test case1: single bdev can'\''t be used in multiple subsystems' 00:15:06.001 test case1: single bdev can't be used in multiple subsystems 00:15:06.001 18:00:59 nvmf_tcp.nvmf_nmic -- target/nmic.sh@26 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK2 00:15:06.001 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:06.001 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:15:06.001 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:06.001 18:00:59 nvmf_tcp.nvmf_nmic -- target/nmic.sh@27 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:15:06.001 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:06.001 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:15:06.001 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:06.001 18:00:59 nvmf_tcp.nvmf_nmic -- target/nmic.sh@28 -- # nmic_status=0 00:15:06.001 18:00:59 nvmf_tcp.nvmf_nmic -- target/nmic.sh@29 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Malloc0 00:15:06.001 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:06.001 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:15:06.002 [2024-07-15 18:00:59.502014] bdev.c:8078:bdev_open: *ERROR*: bdev Malloc0 already claimed: type exclusive_write by module NVMe-oF Target 00:15:06.002 [2024-07-15 18:00:59.502031] subsystem.c:2083:spdk_nvmf_subsystem_add_ns_ext: *ERROR*: Subsystem nqn.2016-06.io.spdk:cnode2: bdev Malloc0 cannot be opened, error=-1 00:15:06.002 [2024-07-15 18:00:59.502038] nvmf_rpc.c:1546:nvmf_rpc_ns_paused: *ERROR*: Unable to add namespace 00:15:06.002 request: 00:15:06.002 { 00:15:06.002 "nqn": "nqn.2016-06.io.spdk:cnode2", 00:15:06.002 "namespace": { 00:15:06.002 "bdev_name": "Malloc0", 00:15:06.002 "no_auto_visible": false 00:15:06.002 }, 00:15:06.002 "method": "nvmf_subsystem_add_ns", 00:15:06.002 "req_id": 1 00:15:06.002 } 00:15:06.002 Got JSON-RPC error response 00:15:06.002 response: 00:15:06.002 { 00:15:06.002 "code": -32602, 00:15:06.002 "message": "Invalid parameters" 00:15:06.002 } 00:15:06.002 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@587 -- # [[ 1 == 0 ]] 00:15:06.002 18:00:59 nvmf_tcp.nvmf_nmic -- target/nmic.sh@29 -- # nmic_status=1 00:15:06.002 18:00:59 nvmf_tcp.nvmf_nmic -- target/nmic.sh@31 -- # '[' 1 -eq 0 ']' 00:15:06.002 18:00:59 nvmf_tcp.nvmf_nmic -- target/nmic.sh@36 -- # echo ' Adding namespace failed - expected result.' 00:15:06.002 Adding namespace failed - expected result. 00:15:06.002 18:00:59 nvmf_tcp.nvmf_nmic -- target/nmic.sh@39 -- # echo 'test case2: host connect to nvmf target in multiple paths' 00:15:06.002 test case2: host connect to nvmf target in multiple paths 00:15:06.002 18:00:59 nvmf_tcp.nvmf_nmic -- target/nmic.sh@40 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:15:06.002 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:06.002 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:15:06.002 [2024-07-15 18:00:59.514141] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:15:06.002 18:00:59 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:06.002 18:00:59 nvmf_tcp.nvmf_nmic -- target/nmic.sh@41 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:15:07.379 18:01:00 nvmf_tcp.nvmf_nmic -- target/nmic.sh@42 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4421 00:15:08.319 18:01:01 nvmf_tcp.nvmf_nmic -- target/nmic.sh@44 -- # waitforserial SPDKISFASTANDAWESOME 00:15:08.319 18:01:01 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@1198 -- # local i=0 00:15:08.319 18:01:01 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:15:08.319 18:01:01 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@1200 -- # [[ -n '' ]] 00:15:08.319 18:01:01 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@1205 -- # sleep 2 00:15:10.241 18:01:03 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:15:10.241 18:01:03 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:15:10.241 18:01:03 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:15:10.241 18:01:03 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@1207 -- # nvme_devices=1 00:15:10.241 18:01:03 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:15:10.241 18:01:03 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@1208 -- # return 0 00:15:10.241 18:01:03 nvmf_tcp.nvmf_nmic -- target/nmic.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 1 -v 00:15:10.241 [global] 00:15:10.241 thread=1 00:15:10.241 invalidate=1 00:15:10.241 rw=write 00:15:10.241 time_based=1 00:15:10.241 runtime=1 00:15:10.241 ioengine=libaio 00:15:10.241 direct=1 00:15:10.241 bs=4096 00:15:10.241 iodepth=1 00:15:10.241 norandommap=0 00:15:10.241 numjobs=1 00:15:10.241 00:15:10.241 verify_dump=1 00:15:10.241 verify_backlog=512 00:15:10.241 verify_state_save=0 00:15:10.241 do_verify=1 00:15:10.241 verify=crc32c-intel 00:15:10.241 [job0] 00:15:10.241 filename=/dev/nvme0n1 00:15:10.241 Could not set queue depth (nvme0n1) 00:15:10.499 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:15:10.499 fio-3.35 00:15:10.499 Starting 1 thread 00:15:11.875 00:15:11.875 job0: (groupid=0, jobs=1): err= 0: pid=569055: Mon Jul 15 18:01:05 2024 00:15:11.875 read: IOPS=21, BW=87.6KiB/s (89.8kB/s)(88.0KiB/1004msec) 00:15:11.875 slat (nsec): min=9353, max=23143, avg=21756.23, stdev=2786.99 00:15:11.875 clat (usec): min=40779, max=42008, avg=41530.57, stdev=470.04 00:15:11.875 lat (usec): min=40801, max=42031, avg=41552.32, stdev=470.21 00:15:11.875 clat percentiles (usec): 00:15:11.875 | 1.00th=[40633], 5.00th=[41157], 10.00th=[41157], 20.00th=[41157], 00:15:11.875 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41681], 60.00th=[41681], 00:15:11.875 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:15:11.875 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42206], 99.95th=[42206], 00:15:11.875 | 99.99th=[42206] 00:15:11.875 write: IOPS=509, BW=2040KiB/s (2089kB/s)(2048KiB/1004msec); 0 zone resets 00:15:11.875 slat (nsec): min=8748, max=41646, avg=9933.21, stdev=2014.00 00:15:11.875 clat (usec): min=148, max=395, avg=163.62, stdev=15.39 00:15:11.875 lat (usec): min=158, max=437, avg=173.56, stdev=16.50 00:15:11.875 clat percentiles (usec): 00:15:11.875 | 1.00th=[ 151], 5.00th=[ 155], 10.00th=[ 155], 20.00th=[ 157], 00:15:11.875 | 30.00th=[ 159], 40.00th=[ 159], 50.00th=[ 161], 60.00th=[ 163], 00:15:11.875 | 70.00th=[ 163], 80.00th=[ 165], 90.00th=[ 172], 95.00th=[ 190], 00:15:11.875 | 99.00th=[ 208], 99.50th=[ 210], 99.90th=[ 396], 99.95th=[ 396], 00:15:11.875 | 99.99th=[ 396] 00:15:11.875 bw ( KiB/s): min= 4096, max= 4096, per=100.00%, avg=4096.00, stdev= 0.00, samples=1 00:15:11.875 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:15:11.875 lat (usec) : 250=95.51%, 500=0.37% 00:15:11.875 lat (msec) : 50=4.12% 00:15:11.875 cpu : usr=0.40%, sys=0.40%, ctx=534, majf=0, minf=2 00:15:11.875 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:15:11.875 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:11.876 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:11.876 issued rwts: total=22,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:11.876 latency : target=0, window=0, percentile=100.00%, depth=1 00:15:11.876 00:15:11.876 Run status group 0 (all jobs): 00:15:11.876 READ: bw=87.6KiB/s (89.8kB/s), 87.6KiB/s-87.6KiB/s (89.8kB/s-89.8kB/s), io=88.0KiB (90.1kB), run=1004-1004msec 00:15:11.876 WRITE: bw=2040KiB/s (2089kB/s), 2040KiB/s-2040KiB/s (2089kB/s-2089kB/s), io=2048KiB (2097kB), run=1004-1004msec 00:15:11.876 00:15:11.876 Disk stats (read/write): 00:15:11.876 nvme0n1: ios=69/512, merge=0/0, ticks=897/83, in_queue=980, util=95.59% 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- target/nmic.sh@48 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:15:11.876 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 2 controller(s) 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- target/nmic.sh@49 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@1219 -- # local i=0 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@1231 -- # return 0 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- target/nmic.sh@51 -- # trap - SIGINT SIGTERM EXIT 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- target/nmic.sh@53 -- # nvmftestfini 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@488 -- # nvmfcleanup 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@117 -- # sync 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@120 -- # set +e 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@121 -- # for i in {1..20} 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:15:11.876 rmmod nvme_tcp 00:15:11.876 rmmod nvme_fabrics 00:15:11.876 rmmod nvme_keyring 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@124 -- # set -e 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@125 -- # return 0 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@489 -- # '[' -n 568004 ']' 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@490 -- # killprocess 568004 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@948 -- # '[' -z 568004 ']' 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@952 -- # kill -0 568004 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@953 -- # uname 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 568004 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@966 -- # echo 'killing process with pid 568004' 00:15:11.876 killing process with pid 568004 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@967 -- # kill 568004 00:15:11.876 18:01:05 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@972 -- # wait 568004 00:15:12.134 18:01:05 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:15:12.134 18:01:05 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:15:12.134 18:01:05 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:15:12.134 18:01:05 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:15:12.134 18:01:05 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@278 -- # remove_spdk_ns 00:15:12.134 18:01:05 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:12.134 18:01:05 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:15:12.134 18:01:05 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:14.669 18:01:07 nvmf_tcp.nvmf_nmic -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:15:14.669 00:15:14.669 real 0m14.726s 00:15:14.669 user 0m34.913s 00:15:14.669 sys 0m4.702s 00:15:14.669 18:01:07 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@1124 -- # xtrace_disable 00:15:14.669 18:01:07 nvmf_tcp.nvmf_nmic -- common/autotest_common.sh@10 -- # set +x 00:15:14.669 ************************************ 00:15:14.669 END TEST nvmf_nmic 00:15:14.669 ************************************ 00:15:14.669 18:01:07 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:15:14.669 18:01:07 nvmf_tcp -- nvmf/nvmf.sh@55 -- # run_test nvmf_fio_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fio.sh --transport=tcp 00:15:14.669 18:01:07 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:15:14.669 18:01:07 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:15:14.669 18:01:07 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:15:14.669 ************************************ 00:15:14.669 START TEST nvmf_fio_target 00:15:14.669 ************************************ 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/fio.sh --transport=tcp 00:15:14.669 * Looking for test storage... 00:15:14.669 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- target/fio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@7 -- # uname -s 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- paths/export.sh@5 -- # export PATH 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@47 -- # : 0 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@51 -- # have_pci_nics=0 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- target/fio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- target/fio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- target/fio.sh@14 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- target/fio.sh@16 -- # nvmftestinit 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@448 -- # prepare_net_devs 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@410 -- # local -g is_hw=no 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@412 -- # remove_spdk_ns 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:15:14.669 18:01:07 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:14.669 18:01:08 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:15:14.669 18:01:08 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:15:14.669 18:01:08 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@285 -- # xtrace_disable 00:15:14.669 18:01:08 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@291 -- # pci_devs=() 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@291 -- # local -a pci_devs 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@292 -- # pci_net_devs=() 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@293 -- # pci_drivers=() 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@293 -- # local -A pci_drivers 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@295 -- # net_devs=() 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@295 -- # local -ga net_devs 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@296 -- # e810=() 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@296 -- # local -ga e810 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@297 -- # x722=() 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@297 -- # local -ga x722 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@298 -- # mlx=() 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@298 -- # local -ga mlx 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:15:19.943 Found 0000:86:00.0 (0x8086 - 0x159b) 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:15:19.943 Found 0000:86:00.1 (0x8086 - 0x159b) 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:15:19.943 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@390 -- # [[ up == up ]] 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:15:19.944 Found net devices under 0000:86:00.0: cvl_0_0 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@390 -- # [[ up == up ]] 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:15:19.944 Found net devices under 0000:86:00.1: cvl_0_1 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@414 -- # is_hw=yes 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:15:19.944 18:01:12 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:15:19.944 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:15:19.944 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.181 ms 00:15:19.944 00:15:19.944 --- 10.0.0.2 ping statistics --- 00:15:19.944 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:19.944 rtt min/avg/max/mdev = 0.181/0.181/0.181/0.000 ms 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:15:19.944 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:15:19.944 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.244 ms 00:15:19.944 00:15:19.944 --- 10.0.0.1 ping statistics --- 00:15:19.944 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:19.944 rtt min/avg/max/mdev = 0.244/0.244/0.244/0.000 ms 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@422 -- # return 0 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- target/fio.sh@17 -- # nvmfappstart -m 0xF 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@722 -- # xtrace_disable 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@481 -- # nvmfpid=572769 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@482 -- # waitforlisten 572769 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@829 -- # '[' -z 572769 ']' 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@834 -- # local max_retries=100 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:19.944 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@838 -- # xtrace_disable 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:15:19.944 18:01:13 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:15:19.944 [2024-07-15 18:01:13.214063] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:15:19.944 [2024-07-15 18:01:13.214106] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:19.944 EAL: No free 2048 kB hugepages reported on node 1 00:15:19.944 [2024-07-15 18:01:13.271088] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:15:19.944 [2024-07-15 18:01:13.351935] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:15:19.944 [2024-07-15 18:01:13.351969] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:15:19.944 [2024-07-15 18:01:13.351976] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:15:19.944 [2024-07-15 18:01:13.351982] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:15:19.944 [2024-07-15 18:01:13.351988] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:15:19.944 [2024-07-15 18:01:13.352146] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:15:19.944 [2024-07-15 18:01:13.352231] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:15:19.944 [2024-07-15 18:01:13.352344] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:15:19.944 [2024-07-15 18:01:13.352345] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:15:20.512 18:01:14 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:15:20.512 18:01:14 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@862 -- # return 0 00:15:20.512 18:01:14 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:15:20.512 18:01:14 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@728 -- # xtrace_disable 00:15:20.512 18:01:14 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:15:20.512 18:01:14 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:15:20.512 18:01:14 nvmf_tcp.nvmf_fio_target -- target/fio.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:15:20.512 [2024-07-15 18:01:14.227703] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:15:20.770 18:01:14 nvmf_tcp.nvmf_fio_target -- target/fio.sh@21 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:15:20.770 18:01:14 nvmf_tcp.nvmf_fio_target -- target/fio.sh@21 -- # malloc_bdevs='Malloc0 ' 00:15:20.770 18:01:14 nvmf_tcp.nvmf_fio_target -- target/fio.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:15:21.029 18:01:14 nvmf_tcp.nvmf_fio_target -- target/fio.sh@22 -- # malloc_bdevs+=Malloc1 00:15:21.029 18:01:14 nvmf_tcp.nvmf_fio_target -- target/fio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:15:21.287 18:01:14 nvmf_tcp.nvmf_fio_target -- target/fio.sh@24 -- # raid_malloc_bdevs='Malloc2 ' 00:15:21.288 18:01:14 nvmf_tcp.nvmf_fio_target -- target/fio.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:15:21.545 18:01:15 nvmf_tcp.nvmf_fio_target -- target/fio.sh@25 -- # raid_malloc_bdevs+=Malloc3 00:15:21.545 18:01:15 nvmf_tcp.nvmf_fio_target -- target/fio.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n raid0 -z 64 -r 0 -b 'Malloc2 Malloc3' 00:15:21.545 18:01:15 nvmf_tcp.nvmf_fio_target -- target/fio.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:15:21.803 18:01:15 nvmf_tcp.nvmf_fio_target -- target/fio.sh@29 -- # concat_malloc_bdevs='Malloc4 ' 00:15:21.803 18:01:15 nvmf_tcp.nvmf_fio_target -- target/fio.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:15:22.062 18:01:15 nvmf_tcp.nvmf_fio_target -- target/fio.sh@30 -- # concat_malloc_bdevs+='Malloc5 ' 00:15:22.062 18:01:15 nvmf_tcp.nvmf_fio_target -- target/fio.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:15:22.321 18:01:15 nvmf_tcp.nvmf_fio_target -- target/fio.sh@31 -- # concat_malloc_bdevs+=Malloc6 00:15:22.321 18:01:15 nvmf_tcp.nvmf_fio_target -- target/fio.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_create -n concat0 -r concat -z 64 -b 'Malloc4 Malloc5 Malloc6' 00:15:22.321 18:01:15 nvmf_tcp.nvmf_fio_target -- target/fio.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDKISFASTANDAWESOME 00:15:22.579 18:01:16 nvmf_tcp.nvmf_fio_target -- target/fio.sh@35 -- # for malloc_bdev in $malloc_bdevs 00:15:22.579 18:01:16 nvmf_tcp.nvmf_fio_target -- target/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:15:22.837 18:01:16 nvmf_tcp.nvmf_fio_target -- target/fio.sh@35 -- # for malloc_bdev in $malloc_bdevs 00:15:22.837 18:01:16 nvmf_tcp.nvmf_fio_target -- target/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:15:22.837 18:01:16 nvmf_tcp.nvmf_fio_target -- target/fio.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:15:23.094 [2024-07-15 18:01:16.677206] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:15:23.094 18:01:16 nvmf_tcp.nvmf_fio_target -- target/fio.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 raid0 00:15:23.353 18:01:16 nvmf_tcp.nvmf_fio_target -- target/fio.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 concat0 00:15:23.353 18:01:17 nvmf_tcp.nvmf_fio_target -- target/fio.sh@46 -- # nvme connect --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -t tcp -n nqn.2016-06.io.spdk:cnode1 -a 10.0.0.2 -s 4420 00:15:24.731 18:01:18 nvmf_tcp.nvmf_fio_target -- target/fio.sh@48 -- # waitforserial SPDKISFASTANDAWESOME 4 00:15:24.731 18:01:18 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@1198 -- # local i=0 00:15:24.731 18:01:18 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@1199 -- # local nvme_device_counter=1 nvme_devices=0 00:15:24.731 18:01:18 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@1200 -- # [[ -n 4 ]] 00:15:24.731 18:01:18 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@1201 -- # nvme_device_counter=4 00:15:24.731 18:01:18 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@1205 -- # sleep 2 00:15:26.667 18:01:20 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@1206 -- # (( i++ <= 15 )) 00:15:26.667 18:01:20 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@1207 -- # lsblk -l -o NAME,SERIAL 00:15:26.667 18:01:20 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@1207 -- # grep -c SPDKISFASTANDAWESOME 00:15:26.667 18:01:20 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@1207 -- # nvme_devices=4 00:15:26.667 18:01:20 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@1208 -- # (( nvme_devices == nvme_device_counter )) 00:15:26.667 18:01:20 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@1208 -- # return 0 00:15:26.667 18:01:20 nvmf_tcp.nvmf_fio_target -- target/fio.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t write -r 1 -v 00:15:26.667 [global] 00:15:26.667 thread=1 00:15:26.667 invalidate=1 00:15:26.667 rw=write 00:15:26.667 time_based=1 00:15:26.667 runtime=1 00:15:26.667 ioengine=libaio 00:15:26.667 direct=1 00:15:26.667 bs=4096 00:15:26.667 iodepth=1 00:15:26.667 norandommap=0 00:15:26.667 numjobs=1 00:15:26.667 00:15:26.667 verify_dump=1 00:15:26.667 verify_backlog=512 00:15:26.667 verify_state_save=0 00:15:26.667 do_verify=1 00:15:26.667 verify=crc32c-intel 00:15:26.667 [job0] 00:15:26.667 filename=/dev/nvme0n1 00:15:26.667 [job1] 00:15:26.667 filename=/dev/nvme0n2 00:15:26.667 [job2] 00:15:26.667 filename=/dev/nvme0n3 00:15:26.667 [job3] 00:15:26.667 filename=/dev/nvme0n4 00:15:26.667 Could not set queue depth (nvme0n1) 00:15:26.667 Could not set queue depth (nvme0n2) 00:15:26.667 Could not set queue depth (nvme0n3) 00:15:26.667 Could not set queue depth (nvme0n4) 00:15:26.925 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:15:26.925 job1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:15:26.925 job2: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:15:26.925 job3: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:15:26.925 fio-3.35 00:15:26.925 Starting 4 threads 00:15:28.301 00:15:28.301 job0: (groupid=0, jobs=1): err= 0: pid=574118: Mon Jul 15 18:01:21 2024 00:15:28.301 read: IOPS=1881, BW=7524KiB/s (7705kB/s)(7532KiB/1001msec) 00:15:28.301 slat (nsec): min=6530, max=24635, avg=7242.15, stdev=839.42 00:15:28.301 clat (usec): min=248, max=467, avg=316.55, stdev=39.06 00:15:28.301 lat (usec): min=255, max=474, avg=323.79, stdev=39.07 00:15:28.301 clat percentiles (usec): 00:15:28.301 | 1.00th=[ 273], 5.00th=[ 285], 10.00th=[ 289], 20.00th=[ 293], 00:15:28.301 | 30.00th=[ 297], 40.00th=[ 302], 50.00th=[ 306], 60.00th=[ 310], 00:15:28.301 | 70.00th=[ 314], 80.00th=[ 318], 90.00th=[ 396], 95.00th=[ 424], 00:15:28.301 | 99.00th=[ 449], 99.50th=[ 453], 99.90th=[ 465], 99.95th=[ 469], 00:15:28.301 | 99.99th=[ 469] 00:15:28.301 write: IOPS=2045, BW=8184KiB/s (8380kB/s)(8192KiB/1001msec); 0 zone resets 00:15:28.301 slat (nsec): min=9594, max=41555, avg=10685.00, stdev=1326.98 00:15:28.301 clat (usec): min=141, max=284, avg=173.98, stdev=14.85 00:15:28.301 lat (usec): min=152, max=296, avg=184.66, stdev=14.98 00:15:28.301 clat percentiles (usec): 00:15:28.301 | 1.00th=[ 151], 5.00th=[ 155], 10.00th=[ 157], 20.00th=[ 163], 00:15:28.301 | 30.00th=[ 165], 40.00th=[ 169], 50.00th=[ 174], 60.00th=[ 176], 00:15:28.301 | 70.00th=[ 180], 80.00th=[ 184], 90.00th=[ 192], 95.00th=[ 202], 00:15:28.301 | 99.00th=[ 223], 99.50th=[ 227], 99.90th=[ 255], 99.95th=[ 281], 00:15:28.301 | 99.99th=[ 285] 00:15:28.301 bw ( KiB/s): min= 8192, max= 8192, per=45.32%, avg=8192.00, stdev= 0.00, samples=1 00:15:28.301 iops : min= 2048, max= 2048, avg=2048.00, stdev= 0.00, samples=1 00:15:28.301 lat (usec) : 250=52.05%, 500=47.95% 00:15:28.301 cpu : usr=2.60%, sys=3.30%, ctx=3931, majf=0, minf=1 00:15:28.301 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:15:28.301 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:28.301 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:28.301 issued rwts: total=1883,2048,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:28.301 latency : target=0, window=0, percentile=100.00%, depth=1 00:15:28.301 job1: (groupid=0, jobs=1): err= 0: pid=574119: Mon Jul 15 18:01:21 2024 00:15:28.301 read: IOPS=21, BW=84.9KiB/s (86.9kB/s)(88.0KiB/1037msec) 00:15:28.301 slat (nsec): min=10402, max=25757, avg=22417.82, stdev=3284.92 00:15:28.301 clat (usec): min=40865, max=41208, avg=40972.22, stdev=76.27 00:15:28.301 lat (usec): min=40884, max=41228, avg=40994.64, stdev=76.01 00:15:28.301 clat percentiles (usec): 00:15:28.301 | 1.00th=[40633], 5.00th=[40633], 10.00th=[40633], 20.00th=[41157], 00:15:28.301 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41157], 60.00th=[41157], 00:15:28.301 | 70.00th=[41157], 80.00th=[41157], 90.00th=[41157], 95.00th=[41157], 00:15:28.301 | 99.00th=[41157], 99.50th=[41157], 99.90th=[41157], 99.95th=[41157], 00:15:28.301 | 99.99th=[41157] 00:15:28.301 write: IOPS=493, BW=1975KiB/s (2022kB/s)(2048KiB/1037msec); 0 zone resets 00:15:28.301 slat (nsec): min=10168, max=37765, avg=12323.76, stdev=2107.80 00:15:28.301 clat (usec): min=162, max=439, avg=247.26, stdev=53.08 00:15:28.301 lat (usec): min=174, max=452, avg=259.58, stdev=53.21 00:15:28.301 clat percentiles (usec): 00:15:28.301 | 1.00th=[ 167], 5.00th=[ 174], 10.00th=[ 184], 20.00th=[ 196], 00:15:28.301 | 30.00th=[ 208], 40.00th=[ 231], 50.00th=[ 243], 60.00th=[ 260], 00:15:28.301 | 70.00th=[ 273], 80.00th=[ 285], 90.00th=[ 306], 95.00th=[ 359], 00:15:28.301 | 99.00th=[ 392], 99.50th=[ 396], 99.90th=[ 441], 99.95th=[ 441], 00:15:28.301 | 99.99th=[ 441] 00:15:28.301 bw ( KiB/s): min= 4096, max= 4096, per=22.66%, avg=4096.00, stdev= 0.00, samples=1 00:15:28.301 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:15:28.301 lat (usec) : 250=50.75%, 500=45.13% 00:15:28.301 lat (msec) : 50=4.12% 00:15:28.301 cpu : usr=0.29%, sys=1.06%, ctx=534, majf=0, minf=2 00:15:28.301 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:15:28.301 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:28.301 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:28.301 issued rwts: total=22,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:28.301 latency : target=0, window=0, percentile=100.00%, depth=1 00:15:28.301 job2: (groupid=0, jobs=1): err= 0: pid=574121: Mon Jul 15 18:01:21 2024 00:15:28.301 read: IOPS=1534, BW=6138KiB/s (6285kB/s)(6144KiB/1001msec) 00:15:28.301 slat (usec): min=7, max=121, avg= 9.70, stdev= 4.34 00:15:28.301 clat (usec): min=304, max=641, avg=378.08, stdev=53.74 00:15:28.301 lat (usec): min=312, max=726, avg=387.78, stdev=55.40 00:15:28.301 clat percentiles (usec): 00:15:28.301 | 1.00th=[ 318], 5.00th=[ 330], 10.00th=[ 338], 20.00th=[ 343], 00:15:28.301 | 30.00th=[ 351], 40.00th=[ 355], 50.00th=[ 359], 60.00th=[ 363], 00:15:28.301 | 70.00th=[ 375], 80.00th=[ 404], 90.00th=[ 461], 95.00th=[ 486], 00:15:28.301 | 99.00th=[ 578], 99.50th=[ 603], 99.90th=[ 627], 99.95th=[ 644], 00:15:28.301 | 99.99th=[ 644] 00:15:28.301 write: IOPS=1612, BW=6450KiB/s (6604kB/s)(6456KiB/1001msec); 0 zone resets 00:15:28.301 slat (usec): min=11, max=17995, avg=24.55, stdev=447.60 00:15:28.301 clat (usec): min=164, max=433, avg=219.50, stdev=33.90 00:15:28.301 lat (usec): min=176, max=18332, avg=244.05, stdev=451.83 00:15:28.301 clat percentiles (usec): 00:15:28.301 | 1.00th=[ 178], 5.00th=[ 186], 10.00th=[ 192], 20.00th=[ 196], 00:15:28.301 | 30.00th=[ 202], 40.00th=[ 204], 50.00th=[ 208], 60.00th=[ 215], 00:15:28.301 | 70.00th=[ 221], 80.00th=[ 239], 90.00th=[ 269], 95.00th=[ 289], 00:15:28.301 | 99.00th=[ 338], 99.50th=[ 351], 99.90th=[ 412], 99.95th=[ 433], 00:15:28.301 | 99.99th=[ 433] 00:15:28.301 bw ( KiB/s): min= 8192, max= 8192, per=45.32%, avg=8192.00, stdev= 0.00, samples=1 00:15:28.301 iops : min= 2048, max= 2048, avg=2048.00, stdev= 0.00, samples=1 00:15:28.301 lat (usec) : 250=43.11%, 500=55.05%, 750=1.84% 00:15:28.301 cpu : usr=2.70%, sys=5.50%, ctx=3152, majf=0, minf=1 00:15:28.301 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:15:28.301 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:28.301 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:28.301 issued rwts: total=1536,1614,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:28.301 latency : target=0, window=0, percentile=100.00%, depth=1 00:15:28.301 job3: (groupid=0, jobs=1): err= 0: pid=574122: Mon Jul 15 18:01:21 2024 00:15:28.301 read: IOPS=19, BW=78.8KiB/s (80.7kB/s)(80.0KiB/1015msec) 00:15:28.301 slat (nsec): min=8464, max=23617, avg=16472.90, stdev=6342.78 00:15:28.301 clat (usec): min=40896, max=41927, avg=41029.21, stdev=218.34 00:15:28.301 lat (usec): min=40911, max=41937, avg=41045.68, stdev=216.75 00:15:28.301 clat percentiles (usec): 00:15:28.301 | 1.00th=[41157], 5.00th=[41157], 10.00th=[41157], 20.00th=[41157], 00:15:28.301 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41157], 60.00th=[41157], 00:15:28.301 | 70.00th=[41157], 80.00th=[41157], 90.00th=[41157], 95.00th=[41157], 00:15:28.301 | 99.00th=[41681], 99.50th=[41681], 99.90th=[41681], 99.95th=[41681], 00:15:28.301 | 99.99th=[41681] 00:15:28.301 write: IOPS=504, BW=2018KiB/s (2066kB/s)(2048KiB/1015msec); 0 zone resets 00:15:28.301 slat (usec): min=9, max=41561, avg=131.26, stdev=2034.65 00:15:28.301 clat (usec): min=174, max=422, avg=243.71, stdev=44.79 00:15:28.301 lat (usec): min=185, max=41978, avg=374.96, stdev=2045.48 00:15:28.301 clat percentiles (usec): 00:15:28.301 | 1.00th=[ 182], 5.00th=[ 188], 10.00th=[ 194], 20.00th=[ 202], 00:15:28.301 | 30.00th=[ 210], 40.00th=[ 221], 50.00th=[ 239], 60.00th=[ 255], 00:15:28.301 | 70.00th=[ 269], 80.00th=[ 281], 90.00th=[ 302], 95.00th=[ 330], 00:15:28.301 | 99.00th=[ 363], 99.50th=[ 404], 99.90th=[ 424], 99.95th=[ 424], 00:15:28.301 | 99.99th=[ 424] 00:15:28.301 bw ( KiB/s): min= 4096, max= 4096, per=22.66%, avg=4096.00, stdev= 0.00, samples=1 00:15:28.301 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:15:28.301 lat (usec) : 250=54.70%, 500=41.54% 00:15:28.301 lat (msec) : 50=3.76% 00:15:28.301 cpu : usr=0.30%, sys=0.59%, ctx=537, majf=0, minf=1 00:15:28.301 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:15:28.301 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:28.301 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:28.301 issued rwts: total=20,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:28.301 latency : target=0, window=0, percentile=100.00%, depth=1 00:15:28.301 00:15:28.301 Run status group 0 (all jobs): 00:15:28.301 READ: bw=13.0MiB/s (13.7MB/s), 78.8KiB/s-7524KiB/s (80.7kB/s-7705kB/s), io=13.5MiB (14.2MB), run=1001-1037msec 00:15:28.302 WRITE: bw=17.7MiB/s (18.5MB/s), 1975KiB/s-8184KiB/s (2022kB/s-8380kB/s), io=18.3MiB (19.2MB), run=1001-1037msec 00:15:28.302 00:15:28.302 Disk stats (read/write): 00:15:28.302 nvme0n1: ios=1585/1561, merge=0/0, ticks=515/265, in_queue=780, util=81.96% 00:15:28.302 nvme0n2: ios=66/512, merge=0/0, ticks=727/123, in_queue=850, util=85.57% 00:15:28.302 nvme0n3: ios=1077/1536, merge=0/0, ticks=1261/323, in_queue=1584, util=94.98% 00:15:28.302 nvme0n4: ios=37/512, merge=0/0, ticks=1445/120, in_queue=1565, util=99.56% 00:15:28.302 18:01:21 nvmf_tcp.nvmf_fio_target -- target/fio.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t randwrite -r 1 -v 00:15:28.302 [global] 00:15:28.302 thread=1 00:15:28.302 invalidate=1 00:15:28.302 rw=randwrite 00:15:28.302 time_based=1 00:15:28.302 runtime=1 00:15:28.302 ioengine=libaio 00:15:28.302 direct=1 00:15:28.302 bs=4096 00:15:28.302 iodepth=1 00:15:28.302 norandommap=0 00:15:28.302 numjobs=1 00:15:28.302 00:15:28.302 verify_dump=1 00:15:28.302 verify_backlog=512 00:15:28.302 verify_state_save=0 00:15:28.302 do_verify=1 00:15:28.302 verify=crc32c-intel 00:15:28.302 [job0] 00:15:28.302 filename=/dev/nvme0n1 00:15:28.302 [job1] 00:15:28.302 filename=/dev/nvme0n2 00:15:28.302 [job2] 00:15:28.302 filename=/dev/nvme0n3 00:15:28.302 [job3] 00:15:28.302 filename=/dev/nvme0n4 00:15:28.302 Could not set queue depth (nvme0n1) 00:15:28.302 Could not set queue depth (nvme0n2) 00:15:28.302 Could not set queue depth (nvme0n3) 00:15:28.302 Could not set queue depth (nvme0n4) 00:15:28.559 job0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:15:28.559 job1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:15:28.559 job2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:15:28.559 job3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:15:28.559 fio-3.35 00:15:28.559 Starting 4 threads 00:15:29.934 00:15:29.934 job0: (groupid=0, jobs=1): err= 0: pid=574500: Mon Jul 15 18:01:23 2024 00:15:29.934 read: IOPS=1395, BW=5582KiB/s (5716kB/s)(5588KiB/1001msec) 00:15:29.934 slat (nsec): min=7224, max=23824, avg=8290.40, stdev=1227.67 00:15:29.934 clat (usec): min=218, max=41351, avg=479.87, stdev=2439.18 00:15:29.934 lat (usec): min=226, max=41359, avg=488.16, stdev=2439.37 00:15:29.934 clat percentiles (usec): 00:15:29.934 | 1.00th=[ 249], 5.00th=[ 269], 10.00th=[ 277], 20.00th=[ 289], 00:15:29.934 | 30.00th=[ 306], 40.00th=[ 318], 50.00th=[ 322], 60.00th=[ 330], 00:15:29.934 | 70.00th=[ 343], 80.00th=[ 371], 90.00th=[ 416], 95.00th=[ 453], 00:15:29.934 | 99.00th=[ 502], 99.50th=[ 578], 99.90th=[41157], 99.95th=[41157], 00:15:29.934 | 99.99th=[41157] 00:15:29.934 write: IOPS=1534, BW=6138KiB/s (6285kB/s)(6144KiB/1001msec); 0 zone resets 00:15:29.934 slat (nsec): min=10275, max=37234, avg=11759.10, stdev=1804.33 00:15:29.934 clat (usec): min=153, max=630, avg=189.34, stdev=26.02 00:15:29.934 lat (usec): min=165, max=641, avg=201.10, stdev=26.21 00:15:29.934 clat percentiles (usec): 00:15:29.934 | 1.00th=[ 161], 5.00th=[ 167], 10.00th=[ 172], 20.00th=[ 176], 00:15:29.934 | 30.00th=[ 178], 40.00th=[ 182], 50.00th=[ 184], 60.00th=[ 188], 00:15:29.934 | 70.00th=[ 192], 80.00th=[ 196], 90.00th=[ 208], 95.00th=[ 241], 00:15:29.934 | 99.00th=[ 285], 99.50th=[ 306], 99.90th=[ 400], 99.95th=[ 627], 00:15:29.934 | 99.99th=[ 627] 00:15:29.934 bw ( KiB/s): min= 8192, max= 8192, per=41.20%, avg=8192.00, stdev= 0.00, samples=1 00:15:29.934 iops : min= 2048, max= 2048, avg=2048.00, stdev= 0.00, samples=1 00:15:29.934 lat (usec) : 250=51.52%, 500=47.90%, 750=0.41% 00:15:29.934 lat (msec) : 50=0.17% 00:15:29.934 cpu : usr=4.80%, sys=2.40%, ctx=2934, majf=0, minf=1 00:15:29.934 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:15:29.934 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:29.934 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:29.934 issued rwts: total=1397,1536,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:29.934 latency : target=0, window=0, percentile=100.00%, depth=1 00:15:29.934 job1: (groupid=0, jobs=1): err= 0: pid=574501: Mon Jul 15 18:01:23 2024 00:15:29.934 read: IOPS=38, BW=155KiB/s (159kB/s)(160KiB/1030msec) 00:15:29.934 slat (nsec): min=6354, max=41995, avg=11841.65, stdev=7033.29 00:15:29.934 clat (usec): min=281, max=41250, avg=22665.96, stdev=20467.45 00:15:29.934 lat (usec): min=291, max=41259, avg=22677.80, stdev=20470.67 00:15:29.934 clat percentiles (usec): 00:15:29.934 | 1.00th=[ 281], 5.00th=[ 289], 10.00th=[ 289], 20.00th=[ 302], 00:15:29.934 | 30.00th=[ 330], 40.00th=[ 355], 50.00th=[40633], 60.00th=[41157], 00:15:29.934 | 70.00th=[41157], 80.00th=[41157], 90.00th=[41157], 95.00th=[41157], 00:15:29.934 | 99.00th=[41157], 99.50th=[41157], 99.90th=[41157], 99.95th=[41157], 00:15:29.934 | 99.99th=[41157] 00:15:29.934 write: IOPS=497, BW=1988KiB/s (2036kB/s)(2048KiB/1030msec); 0 zone resets 00:15:29.934 slat (nsec): min=9480, max=68524, avg=12528.47, stdev=4783.48 00:15:29.934 clat (usec): min=173, max=412, avg=223.76, stdev=35.64 00:15:29.934 lat (usec): min=188, max=449, avg=236.29, stdev=36.09 00:15:29.934 clat percentiles (usec): 00:15:29.934 | 1.00th=[ 184], 5.00th=[ 190], 10.00th=[ 196], 20.00th=[ 200], 00:15:29.934 | 30.00th=[ 206], 40.00th=[ 210], 50.00th=[ 215], 60.00th=[ 219], 00:15:29.934 | 70.00th=[ 225], 80.00th=[ 235], 90.00th=[ 277], 95.00th=[ 318], 00:15:29.934 | 99.00th=[ 330], 99.50th=[ 351], 99.90th=[ 412], 99.95th=[ 412], 00:15:29.934 | 99.99th=[ 412] 00:15:29.934 bw ( KiB/s): min= 4096, max= 4096, per=20.60%, avg=4096.00, stdev= 0.00, samples=1 00:15:29.934 iops : min= 1024, max= 1024, avg=1024.00, stdev= 0.00, samples=1 00:15:29.934 lat (usec) : 250=78.80%, 500=17.21% 00:15:29.934 lat (msec) : 50=3.99% 00:15:29.934 cpu : usr=0.39%, sys=0.58%, ctx=555, majf=0, minf=1 00:15:29.934 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:15:29.934 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:29.934 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:29.934 issued rwts: total=40,512,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:29.935 latency : target=0, window=0, percentile=100.00%, depth=1 00:15:29.935 job2: (groupid=0, jobs=1): err= 0: pid=574505: Mon Jul 15 18:01:23 2024 00:15:29.935 read: IOPS=1857, BW=7429KiB/s (7607kB/s)(7436KiB/1001msec) 00:15:29.935 slat (nsec): min=6450, max=37864, avg=8219.61, stdev=1522.38 00:15:29.935 clat (usec): min=245, max=511, avg=289.40, stdev=30.49 00:15:29.935 lat (usec): min=254, max=519, avg=297.62, stdev=30.49 00:15:29.935 clat percentiles (usec): 00:15:29.935 | 1.00th=[ 255], 5.00th=[ 265], 10.00th=[ 269], 20.00th=[ 273], 00:15:29.935 | 30.00th=[ 277], 40.00th=[ 281], 50.00th=[ 285], 60.00th=[ 289], 00:15:29.935 | 70.00th=[ 293], 80.00th=[ 297], 90.00th=[ 310], 95.00th=[ 326], 00:15:29.935 | 99.00th=[ 457], 99.50th=[ 465], 99.90th=[ 506], 99.95th=[ 510], 00:15:29.935 | 99.99th=[ 510] 00:15:29.935 write: IOPS=2045, BW=8184KiB/s (8380kB/s)(8192KiB/1001msec); 0 zone resets 00:15:29.935 slat (nsec): min=9385, max=53643, avg=12070.86, stdev=1945.93 00:15:29.935 clat (usec): min=160, max=490, avg=200.31, stdev=20.74 00:15:29.935 lat (usec): min=171, max=511, avg=212.38, stdev=21.13 00:15:29.935 clat percentiles (usec): 00:15:29.935 | 1.00th=[ 169], 5.00th=[ 178], 10.00th=[ 180], 20.00th=[ 184], 00:15:29.935 | 30.00th=[ 188], 40.00th=[ 192], 50.00th=[ 196], 60.00th=[ 200], 00:15:29.935 | 70.00th=[ 206], 80.00th=[ 215], 90.00th=[ 229], 95.00th=[ 239], 00:15:29.935 | 99.00th=[ 260], 99.50th=[ 281], 99.90th=[ 310], 99.95th=[ 314], 00:15:29.935 | 99.99th=[ 490] 00:15:29.935 bw ( KiB/s): min= 8192, max= 8192, per=41.20%, avg=8192.00, stdev= 0.00, samples=1 00:15:29.935 iops : min= 2048, max= 2048, avg=2048.00, stdev= 0.00, samples=1 00:15:29.935 lat (usec) : 250=51.73%, 500=48.22%, 750=0.05% 00:15:29.935 cpu : usr=2.80%, sys=6.40%, ctx=3908, majf=0, minf=2 00:15:29.935 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:15:29.935 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:29.935 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:29.935 issued rwts: total=1859,2048,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:29.935 latency : target=0, window=0, percentile=100.00%, depth=1 00:15:29.935 job3: (groupid=0, jobs=1): err= 0: pid=574509: Mon Jul 15 18:01:23 2024 00:15:29.935 read: IOPS=999, BW=3996KiB/s (4092kB/s)(4000KiB/1001msec) 00:15:29.935 slat (nsec): min=6455, max=24559, avg=7888.14, stdev=1386.95 00:15:29.935 clat (usec): min=233, max=42097, avg=766.73, stdev=4171.05 00:15:29.935 lat (usec): min=241, max=42104, avg=774.61, stdev=4171.18 00:15:29.935 clat percentiles (usec): 00:15:29.935 | 1.00th=[ 253], 5.00th=[ 273], 10.00th=[ 285], 20.00th=[ 297], 00:15:29.935 | 30.00th=[ 310], 40.00th=[ 318], 50.00th=[ 322], 60.00th=[ 326], 00:15:29.935 | 70.00th=[ 334], 80.00th=[ 343], 90.00th=[ 371], 95.00th=[ 408], 00:15:29.935 | 99.00th=[30278], 99.50th=[41157], 99.90th=[42206], 99.95th=[42206], 00:15:29.935 | 99.99th=[42206] 00:15:29.935 write: IOPS=1022, BW=4092KiB/s (4190kB/s)(4096KiB/1001msec); 0 zone resets 00:15:29.935 slat (usec): min=8, max=102, avg=10.47, stdev= 3.20 00:15:29.935 clat (usec): min=153, max=429, avg=204.87, stdev=33.51 00:15:29.935 lat (usec): min=163, max=531, avg=215.34, stdev=34.36 00:15:29.935 clat percentiles (usec): 00:15:29.935 | 1.00th=[ 161], 5.00th=[ 167], 10.00th=[ 174], 20.00th=[ 180], 00:15:29.935 | 30.00th=[ 184], 40.00th=[ 188], 50.00th=[ 194], 60.00th=[ 198], 00:15:29.935 | 70.00th=[ 217], 80.00th=[ 241], 90.00th=[ 247], 95.00th=[ 273], 00:15:29.935 | 99.00th=[ 285], 99.50th=[ 289], 99.90th=[ 330], 99.95th=[ 429], 00:15:29.935 | 99.99th=[ 429] 00:15:29.935 bw ( KiB/s): min= 5008, max= 5008, per=25.19%, avg=5008.00, stdev= 0.00, samples=1 00:15:29.935 iops : min= 1252, max= 1252, avg=1252.00, stdev= 0.00, samples=1 00:15:29.935 lat (usec) : 250=46.25%, 500=53.01%, 750=0.10% 00:15:29.935 lat (msec) : 2=0.05%, 4=0.05%, 50=0.54% 00:15:29.935 cpu : usr=1.50%, sys=1.40%, ctx=2025, majf=0, minf=1 00:15:29.935 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:15:29.935 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:29.935 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:29.935 issued rwts: total=1000,1024,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:29.935 latency : target=0, window=0, percentile=100.00%, depth=1 00:15:29.935 00:15:29.935 Run status group 0 (all jobs): 00:15:29.935 READ: bw=16.3MiB/s (17.1MB/s), 155KiB/s-7429KiB/s (159kB/s-7607kB/s), io=16.8MiB (17.6MB), run=1001-1030msec 00:15:29.935 WRITE: bw=19.4MiB/s (20.4MB/s), 1988KiB/s-8184KiB/s (2036kB/s-8380kB/s), io=20.0MiB (21.0MB), run=1001-1030msec 00:15:29.935 00:15:29.935 Disk stats (read/write): 00:15:29.935 nvme0n1: ios=1118/1536, merge=0/0, ticks=1603/268, in_queue=1871, util=97.80% 00:15:29.935 nvme0n2: ios=73/512, merge=0/0, ticks=1614/107, in_queue=1721, util=97.16% 00:15:29.935 nvme0n3: ios=1593/1810, merge=0/0, ticks=1002/333, in_queue=1335, util=98.23% 00:15:29.935 nvme0n4: ios=832/1024, merge=0/0, ticks=849/207, in_queue=1056, util=90.88% 00:15:29.935 18:01:23 nvmf_tcp.nvmf_fio_target -- target/fio.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 128 -t write -r 1 -v 00:15:29.935 [global] 00:15:29.935 thread=1 00:15:29.935 invalidate=1 00:15:29.935 rw=write 00:15:29.935 time_based=1 00:15:29.935 runtime=1 00:15:29.935 ioengine=libaio 00:15:29.935 direct=1 00:15:29.935 bs=4096 00:15:29.935 iodepth=128 00:15:29.935 norandommap=0 00:15:29.935 numjobs=1 00:15:29.935 00:15:29.935 verify_dump=1 00:15:29.935 verify_backlog=512 00:15:29.935 verify_state_save=0 00:15:29.935 do_verify=1 00:15:29.935 verify=crc32c-intel 00:15:29.935 [job0] 00:15:29.935 filename=/dev/nvme0n1 00:15:29.935 [job1] 00:15:29.935 filename=/dev/nvme0n2 00:15:29.935 [job2] 00:15:29.935 filename=/dev/nvme0n3 00:15:29.935 [job3] 00:15:29.935 filename=/dev/nvme0n4 00:15:29.935 Could not set queue depth (nvme0n1) 00:15:29.935 Could not set queue depth (nvme0n2) 00:15:29.935 Could not set queue depth (nvme0n3) 00:15:29.935 Could not set queue depth (nvme0n4) 00:15:30.193 job0: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:15:30.193 job1: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:15:30.193 job2: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:15:30.193 job3: (g=0): rw=write, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:15:30.193 fio-3.35 00:15:30.193 Starting 4 threads 00:15:31.571 00:15:31.571 job0: (groupid=0, jobs=1): err= 0: pid=574905: Mon Jul 15 18:01:25 2024 00:15:31.571 read: IOPS=2559, BW=10.00MiB/s (10.5MB/s)(10.0MiB/1005msec) 00:15:31.571 slat (nsec): min=1609, max=20738k, avg=171426.33, stdev=1071942.05 00:15:31.571 clat (usec): min=957, max=59291, avg=19854.17, stdev=8300.07 00:15:31.571 lat (usec): min=4791, max=59299, avg=20025.60, stdev=8377.75 00:15:31.571 clat percentiles (usec): 00:15:31.571 | 1.00th=[ 8291], 5.00th=[10945], 10.00th=[11600], 20.00th=[13566], 00:15:31.571 | 30.00th=[14222], 40.00th=[15664], 50.00th=[18220], 60.00th=[20841], 00:15:31.571 | 70.00th=[23200], 80.00th=[25560], 90.00th=[29230], 95.00th=[33817], 00:15:31.571 | 99.00th=[55313], 99.50th=[58459], 99.90th=[59507], 99.95th=[59507], 00:15:31.571 | 99.99th=[59507] 00:15:31.571 write: IOPS=3056, BW=11.9MiB/s (12.5MB/s)(12.0MiB/1005msec); 0 zone resets 00:15:31.571 slat (usec): min=2, max=12915, avg=177.05, stdev=820.31 00:15:31.571 clat (usec): min=4903, max=59286, avg=24756.16, stdev=12607.83 00:15:31.571 lat (usec): min=4917, max=60741, avg=24933.22, stdev=12680.65 00:15:31.571 clat percentiles (usec): 00:15:31.571 | 1.00th=[ 5342], 5.00th=[ 9241], 10.00th=[10290], 20.00th=[13698], 00:15:31.571 | 30.00th=[19006], 40.00th=[20841], 50.00th=[21365], 60.00th=[21627], 00:15:31.571 | 70.00th=[28181], 80.00th=[38536], 90.00th=[44827], 95.00th=[49546], 00:15:31.571 | 99.00th=[56886], 99.50th=[57934], 99.90th=[57934], 99.95th=[59507], 00:15:31.571 | 99.99th=[59507] 00:15:31.571 bw ( KiB/s): min=10984, max=12664, per=17.24%, avg=11824.00, stdev=1187.94, samples=2 00:15:31.571 iops : min= 2746, max= 3166, avg=2956.00, stdev=296.98, samples=2 00:15:31.571 lat (usec) : 1000=0.02% 00:15:31.571 lat (msec) : 10=5.83%, 20=36.71%, 50=54.64%, 100=2.80% 00:15:31.571 cpu : usr=2.19%, sys=3.88%, ctx=356, majf=0, minf=1 00:15:31.571 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.3%, 32=0.6%, >=64=98.9% 00:15:31.571 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:31.571 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:15:31.571 issued rwts: total=2572,3072,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:31.571 latency : target=0, window=0, percentile=100.00%, depth=128 00:15:31.571 job1: (groupid=0, jobs=1): err= 0: pid=574922: Mon Jul 15 18:01:25 2024 00:15:31.571 read: IOPS=5109, BW=20.0MiB/s (20.9MB/s)(20.0MiB/1002msec) 00:15:31.571 slat (nsec): min=1387, max=10261k, avg=80618.21, stdev=514845.67 00:15:31.571 clat (usec): min=4836, max=65302, avg=10964.57, stdev=5451.68 00:15:31.571 lat (usec): min=4843, max=65307, avg=11045.19, stdev=5506.26 00:15:31.571 clat percentiles (usec): 00:15:31.571 | 1.00th=[ 6587], 5.00th=[ 7767], 10.00th=[ 8225], 20.00th=[ 9110], 00:15:31.571 | 30.00th=[ 9634], 40.00th=[ 9765], 50.00th=[10028], 60.00th=[10290], 00:15:31.571 | 70.00th=[10683], 80.00th=[11338], 90.00th=[12518], 95.00th=[16909], 00:15:31.571 | 99.00th=[45876], 99.50th=[54264], 99.90th=[62653], 99.95th=[65274], 00:15:31.571 | 99.99th=[65274] 00:15:31.571 write: IOPS=5562, BW=21.7MiB/s (22.8MB/s)(21.8MiB/1002msec); 0 zone resets 00:15:31.571 slat (usec): min=2, max=17596, avg=92.47, stdev=528.43 00:15:31.571 clat (usec): min=474, max=65280, avg=12732.72, stdev=9495.79 00:15:31.571 lat (usec): min=2494, max=65285, avg=12825.19, stdev=9557.24 00:15:31.571 clat percentiles (usec): 00:15:31.571 | 1.00th=[ 3294], 5.00th=[ 6915], 10.00th=[ 8586], 20.00th=[ 9503], 00:15:31.571 | 30.00th=[ 9765], 40.00th=[ 9896], 50.00th=[ 9896], 60.00th=[10159], 00:15:31.571 | 70.00th=[10290], 80.00th=[11338], 90.00th=[19530], 95.00th=[41157], 00:15:31.571 | 99.00th=[52167], 99.50th=[56886], 99.90th=[60031], 99.95th=[60031], 00:15:31.571 | 99.99th=[65274] 00:15:31.571 bw ( KiB/s): min=24576, max=24576, per=35.84%, avg=24576.00, stdev= 0.00, samples=1 00:15:31.571 iops : min= 6144, max= 6144, avg=6144.00, stdev= 0.00, samples=1 00:15:31.571 lat (usec) : 500=0.01% 00:15:31.571 lat (msec) : 4=0.86%, 10=50.84%, 20=41.99%, 50=5.03%, 100=1.27% 00:15:31.571 cpu : usr=3.70%, sys=4.40%, ctx=574, majf=0, minf=1 00:15:31.571 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.3%, >=64=99.4% 00:15:31.571 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:31.571 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:15:31.571 issued rwts: total=5120,5574,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:31.571 latency : target=0, window=0, percentile=100.00%, depth=128 00:15:31.571 job2: (groupid=0, jobs=1): err= 0: pid=574957: Mon Jul 15 18:01:25 2024 00:15:31.571 read: IOPS=4072, BW=15.9MiB/s (16.7MB/s)(16.0MiB/1005msec) 00:15:31.571 slat (nsec): min=1084, max=11805k, avg=110250.50, stdev=729662.14 00:15:31.571 clat (usec): min=2148, max=39377, avg=13492.03, stdev=5946.01 00:15:31.571 lat (usec): min=2155, max=39381, avg=13602.28, stdev=6000.34 00:15:31.571 clat percentiles (usec): 00:15:31.571 | 1.00th=[ 2868], 5.00th=[ 3621], 10.00th=[ 6587], 20.00th=[ 9503], 00:15:31.571 | 30.00th=[ 9896], 40.00th=[11731], 50.00th=[12911], 60.00th=[13960], 00:15:31.571 | 70.00th=[15139], 80.00th=[16450], 90.00th=[21103], 95.00th=[25035], 00:15:31.571 | 99.00th=[31065], 99.50th=[34866], 99.90th=[39584], 99.95th=[39584], 00:15:31.571 | 99.99th=[39584] 00:15:31.571 write: IOPS=4075, BW=15.9MiB/s (16.7MB/s)(16.0MiB/1005msec); 0 zone resets 00:15:31.571 slat (nsec): min=1890, max=13465k, avg=124946.08, stdev=672594.87 00:15:31.571 clat (usec): min=241, max=58395, avg=17558.75, stdev=10311.06 00:15:31.571 lat (usec): min=254, max=58406, avg=17683.70, stdev=10367.97 00:15:31.571 clat percentiles (usec): 00:15:31.571 | 1.00th=[ 1762], 5.00th=[ 6587], 10.00th=[ 7701], 20.00th=[ 9241], 00:15:31.571 | 30.00th=[ 9765], 40.00th=[11600], 50.00th=[17171], 60.00th=[19792], 00:15:31.571 | 70.00th=[21365], 80.00th=[21890], 90.00th=[32900], 95.00th=[39584], 00:15:31.571 | 99.00th=[51643], 99.50th=[54789], 99.90th=[57410], 99.95th=[58459], 00:15:31.571 | 99.99th=[58459] 00:15:31.571 bw ( KiB/s): min=15704, max=17064, per=23.89%, avg=16384.00, stdev=961.67, samples=2 00:15:31.571 iops : min= 3926, max= 4266, avg=4096.00, stdev=240.42, samples=2 00:15:31.571 lat (usec) : 250=0.01%, 750=0.09%, 1000=0.04% 00:15:31.571 lat (msec) : 2=0.38%, 4=3.80%, 10=27.72%, 20=41.53%, 50=25.75% 00:15:31.571 lat (msec) : 100=0.68% 00:15:31.571 cpu : usr=2.79%, sys=4.28%, ctx=476, majf=0, minf=1 00:15:31.571 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.2% 00:15:31.571 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:31.571 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:15:31.571 issued rwts: total=4093,4096,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:31.571 latency : target=0, window=0, percentile=100.00%, depth=128 00:15:31.571 job3: (groupid=0, jobs=1): err= 0: pid=574969: Mon Jul 15 18:01:25 2024 00:15:31.571 read: IOPS=4704, BW=18.4MiB/s (19.3MB/s)(19.1MiB/1042msec) 00:15:31.571 slat (nsec): min=1042, max=7664.0k, avg=92014.66, stdev=545403.47 00:15:31.571 clat (usec): min=3068, max=67183, avg=13049.15, stdev=7822.20 00:15:31.571 lat (usec): min=3072, max=67190, avg=13141.16, stdev=7830.55 00:15:31.571 clat percentiles (usec): 00:15:31.571 | 1.00th=[ 3458], 5.00th=[ 6325], 10.00th=[ 8586], 20.00th=[10552], 00:15:31.571 | 30.00th=[11207], 40.00th=[11469], 50.00th=[11600], 60.00th=[11731], 00:15:31.571 | 70.00th=[12387], 80.00th=[13698], 90.00th=[15270], 95.00th=[19268], 00:15:31.571 | 99.00th=[52167], 99.50th=[57410], 99.90th=[62129], 99.95th=[62129], 00:15:31.571 | 99.99th=[67634] 00:15:31.571 write: IOPS=4913, BW=19.2MiB/s (20.1MB/s)(20.0MiB/1042msec); 0 zone resets 00:15:31.571 slat (nsec): min=1803, max=15044k, avg=102482.58, stdev=638739.38 00:15:31.571 clat (usec): min=1659, max=66166, avg=13227.22, stdev=7337.34 00:15:31.571 lat (usec): min=1671, max=68299, avg=13329.70, stdev=7387.17 00:15:31.571 clat percentiles (usec): 00:15:31.571 | 1.00th=[ 3785], 5.00th=[ 8160], 10.00th=[10552], 20.00th=[11076], 00:15:31.571 | 30.00th=[11207], 40.00th=[11338], 50.00th=[11469], 60.00th=[11600], 00:15:31.571 | 70.00th=[11994], 80.00th=[13435], 90.00th=[16581], 95.00th=[22152], 00:15:31.571 | 99.00th=[54264], 99.50th=[62129], 99.90th=[66323], 99.95th=[66323], 00:15:31.571 | 99.99th=[66323] 00:15:31.571 bw ( KiB/s): min=20432, max=20528, per=29.87%, avg=20480.00, stdev=67.88, samples=2 00:15:31.571 iops : min= 5108, max= 5132, avg=5120.00, stdev=16.97, samples=2 00:15:31.571 lat (msec) : 2=0.07%, 4=1.44%, 10=10.71%, 20=82.46%, 50=4.00% 00:15:31.571 lat (msec) : 100=1.33% 00:15:31.571 cpu : usr=2.40%, sys=4.90%, ctx=518, majf=0, minf=1 00:15:31.571 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.3%, >=64=99.4% 00:15:31.571 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:31.571 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:15:31.571 issued rwts: total=4902,5120,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:31.571 latency : target=0, window=0, percentile=100.00%, depth=128 00:15:31.571 00:15:31.571 Run status group 0 (all jobs): 00:15:31.571 READ: bw=62.6MiB/s (65.6MB/s), 10.00MiB/s-20.0MiB/s (10.5MB/s-20.9MB/s), io=65.2MiB (68.3MB), run=1002-1042msec 00:15:31.571 WRITE: bw=67.0MiB/s (70.2MB/s), 11.9MiB/s-21.7MiB/s (12.5MB/s-22.8MB/s), io=69.8MiB (73.2MB), run=1002-1042msec 00:15:31.571 00:15:31.571 Disk stats (read/write): 00:15:31.571 nvme0n1: ios=2089/2471, merge=0/0, ticks=21341/28905, in_queue=50246, util=100.00% 00:15:31.571 nvme0n2: ios=4050/4096, merge=0/0, ticks=31769/39117, in_queue=70886, util=95.47% 00:15:31.571 nvme0n3: ios=3273/3584, merge=0/0, ticks=30314/48749, in_queue=79063, util=96.84% 00:15:31.571 nvme0n4: ios=3751/4096, merge=0/0, ticks=21093/26734, in_queue=47827, util=85.90% 00:15:31.571 18:01:25 nvmf_tcp.nvmf_fio_target -- target/fio.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 128 -t randwrite -r 1 -v 00:15:31.571 [global] 00:15:31.571 thread=1 00:15:31.571 invalidate=1 00:15:31.572 rw=randwrite 00:15:31.572 time_based=1 00:15:31.572 runtime=1 00:15:31.572 ioengine=libaio 00:15:31.572 direct=1 00:15:31.572 bs=4096 00:15:31.572 iodepth=128 00:15:31.572 norandommap=0 00:15:31.572 numjobs=1 00:15:31.572 00:15:31.572 verify_dump=1 00:15:31.572 verify_backlog=512 00:15:31.572 verify_state_save=0 00:15:31.572 do_verify=1 00:15:31.572 verify=crc32c-intel 00:15:31.572 [job0] 00:15:31.572 filename=/dev/nvme0n1 00:15:31.572 [job1] 00:15:31.572 filename=/dev/nvme0n2 00:15:31.572 [job2] 00:15:31.572 filename=/dev/nvme0n3 00:15:31.572 [job3] 00:15:31.572 filename=/dev/nvme0n4 00:15:31.572 Could not set queue depth (nvme0n1) 00:15:31.572 Could not set queue depth (nvme0n2) 00:15:31.572 Could not set queue depth (nvme0n3) 00:15:31.572 Could not set queue depth (nvme0n4) 00:15:31.829 job0: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:15:31.829 job1: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:15:31.829 job2: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:15:31.829 job3: (g=0): rw=randwrite, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=128 00:15:31.829 fio-3.35 00:15:31.829 Starting 4 threads 00:15:33.249 00:15:33.249 job0: (groupid=0, jobs=1): err= 0: pid=575412: Mon Jul 15 18:01:26 2024 00:15:33.249 read: IOPS=4135, BW=16.2MiB/s (16.9MB/s)(16.2MiB/1006msec) 00:15:33.249 slat (nsec): min=1032, max=43856k, avg=106629.80, stdev=863778.00 00:15:33.249 clat (usec): min=2628, max=55879, avg=13067.58, stdev=6543.50 00:15:33.249 lat (usec): min=7279, max=55889, avg=13174.21, stdev=6571.40 00:15:33.249 clat percentiles (usec): 00:15:33.249 | 1.00th=[ 7373], 5.00th=[ 8356], 10.00th=[ 9896], 20.00th=[10814], 00:15:33.249 | 30.00th=[11600], 40.00th=[11994], 50.00th=[12125], 60.00th=[12518], 00:15:33.249 | 70.00th=[12780], 80.00th=[13304], 90.00th=[14484], 95.00th=[16909], 00:15:33.249 | 99.00th=[53740], 99.50th=[54264], 99.90th=[55313], 99.95th=[55837], 00:15:33.249 | 99.99th=[55837] 00:15:33.249 write: IOPS=4580, BW=17.9MiB/s (18.8MB/s)(18.0MiB/1006msec); 0 zone resets 00:15:33.249 slat (nsec): min=1825, max=40237k, avg=116708.11, stdev=971005.31 00:15:33.249 clat (usec): min=5914, max=60373, avg=15782.72, stdev=11656.66 00:15:33.249 lat (usec): min=6171, max=60381, avg=15899.42, stdev=11703.42 00:15:33.249 clat percentiles (usec): 00:15:33.249 | 1.00th=[ 7046], 5.00th=[ 8848], 10.00th=[ 9372], 20.00th=[10421], 00:15:33.249 | 30.00th=[11076], 40.00th=[11469], 50.00th=[11600], 60.00th=[12125], 00:15:33.249 | 70.00th=[12518], 80.00th=[14091], 90.00th=[31589], 95.00th=[50070], 00:15:33.249 | 99.00th=[57410], 99.50th=[57410], 99.90th=[60031], 99.95th=[60556], 00:15:33.249 | 99.99th=[60556] 00:15:33.249 bw ( KiB/s): min=16552, max=19800, per=22.65%, avg=18176.00, stdev=2296.68, samples=2 00:15:33.249 iops : min= 4138, max= 4950, avg=4544.00, stdev=574.17, samples=2 00:15:33.249 lat (msec) : 4=0.01%, 10=13.72%, 20=76.98%, 50=5.44%, 100=3.84% 00:15:33.249 cpu : usr=2.69%, sys=3.68%, ctx=475, majf=0, minf=1 00:15:33.249 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.3% 00:15:33.249 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:33.249 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:15:33.249 issued rwts: total=4160,4608,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:33.249 latency : target=0, window=0, percentile=100.00%, depth=128 00:15:33.249 job1: (groupid=0, jobs=1): err= 0: pid=575424: Mon Jul 15 18:01:26 2024 00:15:33.249 read: IOPS=5104, BW=19.9MiB/s (20.9MB/s)(20.0MiB/1003msec) 00:15:33.249 slat (nsec): min=979, max=17567k, avg=87588.15, stdev=575089.70 00:15:33.249 clat (usec): min=3602, max=60032, avg=12285.67, stdev=5107.73 00:15:33.249 lat (usec): min=3604, max=60036, avg=12373.26, stdev=5107.56 00:15:33.249 clat percentiles (usec): 00:15:33.249 | 1.00th=[ 5145], 5.00th=[ 8455], 10.00th=[ 9241], 20.00th=[10028], 00:15:33.249 | 30.00th=[10552], 40.00th=[10814], 50.00th=[11469], 60.00th=[11731], 00:15:33.249 | 70.00th=[11994], 80.00th=[12387], 90.00th=[14746], 95.00th=[22938], 00:15:33.249 | 99.00th=[38011], 99.50th=[38011], 99.90th=[42206], 99.95th=[42206], 00:15:33.249 | 99.99th=[60031] 00:15:33.249 write: IOPS=5591, BW=21.8MiB/s (22.9MB/s)(21.9MiB/1003msec); 0 zone resets 00:15:33.249 slat (nsec): min=1683, max=11780k, avg=86125.93, stdev=478662.32 00:15:33.249 clat (usec): min=415, max=28959, avg=11246.52, stdev=3019.47 00:15:33.249 lat (usec): min=1066, max=28985, avg=11332.64, stdev=3042.55 00:15:33.249 clat percentiles (usec): 00:15:33.249 | 1.00th=[ 2737], 5.00th=[ 6194], 10.00th=[ 8160], 20.00th=[ 9765], 00:15:33.249 | 30.00th=[10552], 40.00th=[11076], 50.00th=[11469], 60.00th=[11731], 00:15:33.249 | 70.00th=[11863], 80.00th=[11994], 90.00th=[13698], 95.00th=[16909], 00:15:33.249 | 99.00th=[22414], 99.50th=[22414], 99.90th=[22414], 99.95th=[22414], 00:15:33.249 | 99.99th=[28967] 00:15:33.249 bw ( KiB/s): min=20480, max=23368, per=27.33%, avg=21924.00, stdev=2042.12, samples=2 00:15:33.249 iops : min= 5120, max= 5842, avg=5481.00, stdev=510.53, samples=2 00:15:33.249 lat (usec) : 500=0.01% 00:15:33.249 lat (msec) : 2=0.46%, 4=0.92%, 10=18.50%, 20=76.48%, 50=3.62% 00:15:33.249 lat (msec) : 100=0.01% 00:15:33.249 cpu : usr=3.19%, sys=3.89%, ctx=563, majf=0, minf=1 00:15:33.249 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.3%, >=64=99.4% 00:15:33.249 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:33.249 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:15:33.249 issued rwts: total=5120,5608,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:33.249 latency : target=0, window=0, percentile=100.00%, depth=128 00:15:33.249 job2: (groupid=0, jobs=1): err= 0: pid=575441: Mon Jul 15 18:01:26 2024 00:15:33.249 read: IOPS=5074, BW=19.8MiB/s (20.8MB/s)(20.0MiB/1009msec) 00:15:33.249 slat (nsec): min=1284, max=11642k, avg=109116.43, stdev=777794.40 00:15:33.249 clat (usec): min=3750, max=28016, avg=13165.88, stdev=3455.98 00:15:33.249 lat (usec): min=3756, max=28032, avg=13274.99, stdev=3504.60 00:15:33.249 clat percentiles (usec): 00:15:33.249 | 1.00th=[ 4686], 5.00th=[ 8848], 10.00th=[10159], 20.00th=[10814], 00:15:33.249 | 30.00th=[10945], 40.00th=[11469], 50.00th=[12780], 60.00th=[13042], 00:15:33.249 | 70.00th=[13566], 80.00th=[16188], 90.00th=[18220], 95.00th=[20055], 00:15:33.249 | 99.00th=[22414], 99.50th=[22938], 99.90th=[24511], 99.95th=[25035], 00:15:33.249 | 99.99th=[27919] 00:15:33.249 write: IOPS=5365, BW=21.0MiB/s (22.0MB/s)(21.1MiB/1009msec); 0 zone resets 00:15:33.249 slat (usec): min=2, max=10397, avg=76.92, stdev=338.01 00:15:33.249 clat (usec): min=1569, max=23990, avg=11189.69, stdev=2653.68 00:15:33.249 lat (usec): min=1584, max=23993, avg=11266.61, stdev=2684.58 00:15:33.249 clat percentiles (usec): 00:15:33.249 | 1.00th=[ 3064], 5.00th=[ 4948], 10.00th=[ 7177], 20.00th=[ 9896], 00:15:33.249 | 30.00th=[11207], 40.00th=[11338], 50.00th=[11731], 60.00th=[12387], 00:15:33.249 | 70.00th=[12780], 80.00th=[13042], 90.00th=[13173], 95.00th=[13304], 00:15:33.249 | 99.00th=[16909], 99.50th=[19792], 99.90th=[22938], 99.95th=[23725], 00:15:33.249 | 99.99th=[23987] 00:15:33.249 bw ( KiB/s): min=20480, max=21816, per=26.36%, avg=21148.00, stdev=944.69, samples=2 00:15:33.249 iops : min= 5120, max= 5454, avg=5287.00, stdev=236.17, samples=2 00:15:33.249 lat (msec) : 2=0.02%, 4=1.43%, 10=13.93%, 20=81.97%, 50=2.65% 00:15:33.249 cpu : usr=3.77%, sys=4.17%, ctx=708, majf=0, minf=1 00:15:33.249 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.3%, >=64=99.4% 00:15:33.249 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:33.249 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:15:33.249 issued rwts: total=5120,5414,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:33.249 latency : target=0, window=0, percentile=100.00%, depth=128 00:15:33.249 job3: (groupid=0, jobs=1): err= 0: pid=575446: Mon Jul 15 18:01:26 2024 00:15:33.249 read: IOPS=4070, BW=15.9MiB/s (16.7MB/s)(16.0MiB/1008msec) 00:15:33.249 slat (nsec): min=1491, max=8504.8k, avg=102101.27, stdev=574112.23 00:15:33.249 clat (usec): min=5329, max=29535, avg=12526.81, stdev=2530.99 00:15:33.249 lat (usec): min=7493, max=29549, avg=12628.91, stdev=2575.02 00:15:33.249 clat percentiles (usec): 00:15:33.249 | 1.00th=[ 7963], 5.00th=[ 9372], 10.00th=[10159], 20.00th=[10814], 00:15:33.249 | 30.00th=[11207], 40.00th=[11731], 50.00th=[12256], 60.00th=[12518], 00:15:33.249 | 70.00th=[13173], 80.00th=[14091], 90.00th=[14746], 95.00th=[16188], 00:15:33.249 | 99.00th=[23200], 99.50th=[26608], 99.90th=[26608], 99.95th=[26608], 00:15:33.249 | 99.99th=[29492] 00:15:33.249 write: IOPS=4571, BW=17.9MiB/s (18.7MB/s)(18.0MiB/1008msec); 0 zone resets 00:15:33.249 slat (usec): min=2, max=9048, avg=120.41, stdev=596.57 00:15:33.249 clat (usec): min=6265, max=57325, avg=16469.03, stdev=8735.36 00:15:33.249 lat (usec): min=6271, max=57328, avg=16589.44, stdev=8789.19 00:15:33.249 clat percentiles (usec): 00:15:33.249 | 1.00th=[ 7832], 5.00th=[10683], 10.00th=[11076], 20.00th=[11994], 00:15:33.249 | 30.00th=[12387], 40.00th=[12649], 50.00th=[12649], 60.00th=[12911], 00:15:33.249 | 70.00th=[14353], 80.00th=[19792], 90.00th=[30278], 95.00th=[37487], 00:15:33.249 | 99.00th=[48497], 99.50th=[54789], 99.90th=[57410], 99.95th=[57410], 00:15:33.249 | 99.99th=[57410] 00:15:33.249 bw ( KiB/s): min=15168, max=20728, per=22.37%, avg=17948.00, stdev=3931.51, samples=2 00:15:33.249 iops : min= 3792, max= 5182, avg=4487.00, stdev=982.88, samples=2 00:15:33.249 lat (msec) : 10=6.22%, 20=82.17%, 50=11.25%, 100=0.36% 00:15:33.249 cpu : usr=3.97%, sys=4.77%, ctx=498, majf=0, minf=1 00:15:33.249 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.2%, 32=0.4%, >=64=99.3% 00:15:33.249 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:33.249 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:15:33.249 issued rwts: total=4103,4608,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:33.249 latency : target=0, window=0, percentile=100.00%, depth=128 00:15:33.249 00:15:33.249 Run status group 0 (all jobs): 00:15:33.249 READ: bw=71.6MiB/s (75.1MB/s), 15.9MiB/s-19.9MiB/s (16.7MB/s-20.9MB/s), io=72.3MiB (75.8MB), run=1003-1009msec 00:15:33.249 WRITE: bw=78.3MiB/s (82.2MB/s), 17.9MiB/s-21.8MiB/s (18.7MB/s-22.9MB/s), io=79.1MiB (82.9MB), run=1003-1009msec 00:15:33.249 00:15:33.249 Disk stats (read/write): 00:15:33.249 nvme0n1: ios=3607/3895, merge=0/0, ticks=15314/14427, in_queue=29741, util=94.09% 00:15:33.249 nvme0n2: ios=4449/4608, merge=0/0, ticks=29230/19900, in_queue=49130, util=91.48% 00:15:33.249 nvme0n3: ios=4143/4608, merge=0/0, ticks=53475/51424, in_queue=104899, util=93.35% 00:15:33.249 nvme0n4: ios=3794/4096, merge=0/0, ticks=24626/28339, in_queue=52965, util=96.23% 00:15:33.249 18:01:26 nvmf_tcp.nvmf_fio_target -- target/fio.sh@55 -- # sync 00:15:33.249 18:01:26 nvmf_tcp.nvmf_fio_target -- target/fio.sh@59 -- # fio_pid=575527 00:15:33.249 18:01:26 nvmf_tcp.nvmf_fio_target -- target/fio.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/fio-wrapper -p nvmf -i 4096 -d 1 -t read -r 10 00:15:33.249 18:01:26 nvmf_tcp.nvmf_fio_target -- target/fio.sh@61 -- # sleep 3 00:15:33.249 [global] 00:15:33.249 thread=1 00:15:33.249 invalidate=1 00:15:33.249 rw=read 00:15:33.249 time_based=1 00:15:33.249 runtime=10 00:15:33.249 ioengine=libaio 00:15:33.249 direct=1 00:15:33.249 bs=4096 00:15:33.249 iodepth=1 00:15:33.249 norandommap=1 00:15:33.249 numjobs=1 00:15:33.249 00:15:33.249 [job0] 00:15:33.249 filename=/dev/nvme0n1 00:15:33.249 [job1] 00:15:33.249 filename=/dev/nvme0n2 00:15:33.249 [job2] 00:15:33.249 filename=/dev/nvme0n3 00:15:33.250 [job3] 00:15:33.250 filename=/dev/nvme0n4 00:15:33.250 Could not set queue depth (nvme0n1) 00:15:33.250 Could not set queue depth (nvme0n2) 00:15:33.250 Could not set queue depth (nvme0n3) 00:15:33.250 Could not set queue depth (nvme0n4) 00:15:33.509 job0: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:15:33.509 job1: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:15:33.509 job2: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:15:33.509 job3: (g=0): rw=read, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=libaio, iodepth=1 00:15:33.509 fio-3.35 00:15:33.509 Starting 4 threads 00:15:36.026 18:01:29 nvmf_tcp.nvmf_fio_target -- target/fio.sh@63 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_delete concat0 00:15:36.282 18:01:29 nvmf_tcp.nvmf_fio_target -- target/fio.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_raid_delete raid0 00:15:36.282 fio: io_u error on file /dev/nvme0n4: Remote I/O error: read offset=3633152, buflen=4096 00:15:36.282 fio: pid=575831, err=121/file:io_u.c:1889, func=io_u error, error=Remote I/O error 00:15:36.539 fio: io_u error on file /dev/nvme0n3: Remote I/O error: read offset=7491584, buflen=4096 00:15:36.539 fio: pid=575830, err=121/file:io_u.c:1889, func=io_u error, error=Remote I/O error 00:15:36.539 18:01:30 nvmf_tcp.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:15:36.539 18:01:30 nvmf_tcp.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc0 00:15:36.797 18:01:30 nvmf_tcp.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:15:36.797 18:01:30 nvmf_tcp.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc1 00:15:36.797 fio: io_u error on file /dev/nvme0n1: Remote I/O error: read offset=24739840, buflen=4096 00:15:36.797 fio: pid=575828, err=121/file:io_u.c:1889, func=io_u error, error=Remote I/O error 00:15:36.797 fio: io_u error on file /dev/nvme0n2: Remote I/O error: read offset=41017344, buflen=4096 00:15:36.797 fio: pid=575829, err=121/file:io_u.c:1889, func=io_u error, error=Remote I/O error 00:15:36.797 18:01:30 nvmf_tcp.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:15:36.797 18:01:30 nvmf_tcp.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc2 00:15:37.053 00:15:37.054 job0: (groupid=0, jobs=1): err=121 (file:io_u.c:1889, func=io_u error, error=Remote I/O error): pid=575828: Mon Jul 15 18:01:30 2024 00:15:37.054 read: IOPS=1906, BW=7626KiB/s (7809kB/s)(23.6MiB/3168msec) 00:15:37.054 slat (usec): min=2, max=27485, avg=15.88, stdev=431.18 00:15:37.054 clat (usec): min=196, max=42016, avg=502.29, stdev=2840.49 00:15:37.054 lat (usec): min=204, max=42035, avg=518.17, stdev=2873.70 00:15:37.054 clat percentiles (usec): 00:15:37.054 | 1.00th=[ 241], 5.00th=[ 265], 10.00th=[ 269], 20.00th=[ 281], 00:15:37.054 | 30.00th=[ 285], 40.00th=[ 289], 50.00th=[ 297], 60.00th=[ 306], 00:15:37.054 | 70.00th=[ 314], 80.00th=[ 322], 90.00th=[ 338], 95.00th=[ 351], 00:15:37.054 | 99.00th=[ 404], 99.50th=[ 3294], 99.90th=[41157], 99.95th=[41157], 00:15:37.054 | 99.99th=[42206] 00:15:37.054 bw ( KiB/s): min= 104, max=12792, per=34.92%, avg=7881.17, stdev=4922.96, samples=6 00:15:37.054 iops : min= 26, max= 3198, avg=1970.17, stdev=1230.83, samples=6 00:15:37.054 lat (usec) : 250=1.71%, 500=97.60%, 750=0.08% 00:15:37.054 lat (msec) : 2=0.08%, 4=0.02%, 50=0.50% 00:15:37.054 cpu : usr=1.23%, sys=2.90%, ctx=6043, majf=0, minf=1 00:15:37.054 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:15:37.054 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:37.054 complete : 0=0.1%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:37.054 issued rwts: total=6041,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:37.054 latency : target=0, window=0, percentile=100.00%, depth=1 00:15:37.054 job1: (groupid=0, jobs=1): err=121 (file:io_u.c:1889, func=io_u error, error=Remote I/O error): pid=575829: Mon Jul 15 18:01:30 2024 00:15:37.054 read: IOPS=3010, BW=11.8MiB/s (12.3MB/s)(39.1MiB/3327msec) 00:15:37.054 slat (usec): min=7, max=16580, avg=15.37, stdev=318.07 00:15:37.054 clat (usec): min=186, max=42210, avg=312.70, stdev=768.83 00:15:37.054 lat (usec): min=195, max=56839, avg=328.07, stdev=934.67 00:15:37.054 clat percentiles (usec): 00:15:37.054 | 1.00th=[ 229], 5.00th=[ 247], 10.00th=[ 253], 20.00th=[ 260], 00:15:37.054 | 30.00th=[ 265], 40.00th=[ 269], 50.00th=[ 277], 60.00th=[ 285], 00:15:37.054 | 70.00th=[ 310], 80.00th=[ 334], 90.00th=[ 359], 95.00th=[ 400], 00:15:37.054 | 99.00th=[ 461], 99.50th=[ 478], 99.90th=[ 1385], 99.95th=[ 8848], 00:15:37.054 | 99.99th=[42206] 00:15:37.054 bw ( KiB/s): min=10864, max=14784, per=57.02%, avg=12868.50, stdev=1564.20, samples=6 00:15:37.054 iops : min= 2716, max= 3696, avg=3217.00, stdev=391.06, samples=6 00:15:37.054 lat (usec) : 250=7.62%, 500=92.09%, 750=0.14% 00:15:37.054 lat (msec) : 2=0.07%, 4=0.01%, 10=0.01%, 20=0.01%, 50=0.04% 00:15:37.054 cpu : usr=1.95%, sys=4.66%, ctx=10021, majf=0, minf=1 00:15:37.054 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:15:37.054 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:37.054 complete : 0=0.1%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:37.054 issued rwts: total=10015,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:37.054 latency : target=0, window=0, percentile=100.00%, depth=1 00:15:37.054 job2: (groupid=0, jobs=1): err=121 (file:io_u.c:1889, func=io_u error, error=Remote I/O error): pid=575830: Mon Jul 15 18:01:30 2024 00:15:37.054 read: IOPS=623, BW=2494KiB/s (2553kB/s)(7316KiB/2934msec) 00:15:37.054 slat (nsec): min=2273, max=41044, avg=8762.47, stdev=2964.81 00:15:37.054 clat (usec): min=204, max=41565, avg=1581.70, stdev=7079.09 00:15:37.054 lat (usec): min=216, max=41577, avg=1590.47, stdev=7081.16 00:15:37.054 clat percentiles (usec): 00:15:37.054 | 1.00th=[ 219], 5.00th=[ 235], 10.00th=[ 247], 20.00th=[ 269], 00:15:37.054 | 30.00th=[ 277], 40.00th=[ 289], 50.00th=[ 297], 60.00th=[ 318], 00:15:37.054 | 70.00th=[ 334], 80.00th=[ 343], 90.00th=[ 359], 95.00th=[ 408], 00:15:37.054 | 99.00th=[41157], 99.50th=[41157], 99.90th=[41157], 99.95th=[41681], 00:15:37.054 | 99.99th=[41681] 00:15:37.054 bw ( KiB/s): min= 96, max= 3392, per=3.63%, avg=820.80, stdev=1443.90, samples=5 00:15:37.054 iops : min= 24, max= 848, avg=205.20, stdev=360.98, samples=5 00:15:37.054 lat (usec) : 250=11.53%, 500=84.59%, 750=0.44% 00:15:37.054 lat (msec) : 2=0.22%, 20=0.05%, 50=3.11% 00:15:37.054 cpu : usr=0.27%, sys=1.13%, ctx=1831, majf=0, minf=1 00:15:37.054 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:15:37.054 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:37.054 complete : 0=0.1%, 4=99.9%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:37.054 issued rwts: total=1830,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:37.054 latency : target=0, window=0, percentile=100.00%, depth=1 00:15:37.054 job3: (groupid=0, jobs=1): err=121 (file:io_u.c:1889, func=io_u error, error=Remote I/O error): pid=575831: Mon Jul 15 18:01:30 2024 00:15:37.054 read: IOPS=323, BW=1291KiB/s (1322kB/s)(3548KiB/2748msec) 00:15:37.054 slat (nsec): min=6151, max=34015, avg=8142.63, stdev=3610.86 00:15:37.054 clat (usec): min=247, max=42076, avg=3063.53, stdev=10234.04 00:15:37.054 lat (usec): min=254, max=42088, avg=3071.65, stdev=10236.63 00:15:37.054 clat percentiles (usec): 00:15:37.054 | 1.00th=[ 262], 5.00th=[ 269], 10.00th=[ 273], 20.00th=[ 281], 00:15:37.054 | 30.00th=[ 289], 40.00th=[ 293], 50.00th=[ 306], 60.00th=[ 318], 00:15:37.054 | 70.00th=[ 330], 80.00th=[ 343], 90.00th=[ 367], 95.00th=[41157], 00:15:37.054 | 99.00th=[41157], 99.50th=[41681], 99.90th=[42206], 99.95th=[42206], 00:15:37.054 | 99.99th=[42206] 00:15:37.054 bw ( KiB/s): min= 96, max= 3328, per=6.24%, avg=1409.60, stdev=1411.28, samples=5 00:15:37.054 iops : min= 24, max= 832, avg=352.40, stdev=352.82, samples=5 00:15:37.054 lat (usec) : 250=0.11%, 500=92.57%, 750=0.45% 00:15:37.054 lat (msec) : 50=6.76% 00:15:37.054 cpu : usr=0.04%, sys=0.36%, ctx=888, majf=0, minf=2 00:15:37.054 IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:15:37.054 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:37.054 complete : 0=0.1%, 4=99.9%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:15:37.054 issued rwts: total=888,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:15:37.054 latency : target=0, window=0, percentile=100.00%, depth=1 00:15:37.054 00:15:37.054 Run status group 0 (all jobs): 00:15:37.054 READ: bw=22.0MiB/s (23.1MB/s), 1291KiB/s-11.8MiB/s (1322kB/s-12.3MB/s), io=73.3MiB (76.9MB), run=2748-3327msec 00:15:37.054 00:15:37.054 Disk stats (read/write): 00:15:37.054 nvme0n1: ios=6039/0, merge=0/0, ticks=2911/0, in_queue=2911, util=94.36% 00:15:37.054 nvme0n2: ios=10018/0, merge=0/0, ticks=3040/0, in_queue=3040, util=98.18% 00:15:37.054 nvme0n3: ios=1656/0, merge=0/0, ticks=2827/0, in_queue=2827, util=96.52% 00:15:37.054 nvme0n4: ios=884/0, merge=0/0, ticks=2592/0, in_queue=2592, util=96.41% 00:15:37.054 18:01:30 nvmf_tcp.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:15:37.054 18:01:30 nvmf_tcp.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc3 00:15:37.311 18:01:30 nvmf_tcp.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:15:37.311 18:01:30 nvmf_tcp.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc4 00:15:37.567 18:01:31 nvmf_tcp.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:15:37.567 18:01:31 nvmf_tcp.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc5 00:15:37.567 18:01:31 nvmf_tcp.nvmf_fio_target -- target/fio.sh@65 -- # for malloc_bdev in $malloc_bdevs $raid_malloc_bdevs $concat_malloc_bdevs 00:15:37.567 18:01:31 nvmf_tcp.nvmf_fio_target -- target/fio.sh@66 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_delete Malloc6 00:15:37.823 18:01:31 nvmf_tcp.nvmf_fio_target -- target/fio.sh@69 -- # fio_status=0 00:15:37.823 18:01:31 nvmf_tcp.nvmf_fio_target -- target/fio.sh@70 -- # wait 575527 00:15:37.823 18:01:31 nvmf_tcp.nvmf_fio_target -- target/fio.sh@70 -- # fio_status=4 00:15:37.823 18:01:31 nvmf_tcp.nvmf_fio_target -- target/fio.sh@72 -- # nvme disconnect -n nqn.2016-06.io.spdk:cnode1 00:15:37.823 NQN:nqn.2016-06.io.spdk:cnode1 disconnected 1 controller(s) 00:15:37.823 18:01:31 nvmf_tcp.nvmf_fio_target -- target/fio.sh@73 -- # waitforserial_disconnect SPDKISFASTANDAWESOME 00:15:37.823 18:01:31 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@1219 -- # local i=0 00:15:37.823 18:01:31 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@1220 -- # lsblk -o NAME,SERIAL 00:15:37.823 18:01:31 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@1220 -- # grep -q -w SPDKISFASTANDAWESOME 00:15:38.079 18:01:31 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@1227 -- # lsblk -l -o NAME,SERIAL 00:15:38.079 18:01:31 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@1227 -- # grep -q -w SPDKISFASTANDAWESOME 00:15:38.079 18:01:31 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@1231 -- # return 0 00:15:38.079 18:01:31 nvmf_tcp.nvmf_fio_target -- target/fio.sh@75 -- # '[' 4 -eq 0 ']' 00:15:38.079 18:01:31 nvmf_tcp.nvmf_fio_target -- target/fio.sh@80 -- # echo 'nvmf hotplug test: fio failed as expected' 00:15:38.079 nvmf hotplug test: fio failed as expected 00:15:38.079 18:01:31 nvmf_tcp.nvmf_fio_target -- target/fio.sh@83 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:15:38.079 18:01:31 nvmf_tcp.nvmf_fio_target -- target/fio.sh@85 -- # rm -f ./local-job0-0-verify.state 00:15:38.079 18:01:31 nvmf_tcp.nvmf_fio_target -- target/fio.sh@86 -- # rm -f ./local-job1-1-verify.state 00:15:38.079 18:01:31 nvmf_tcp.nvmf_fio_target -- target/fio.sh@87 -- # rm -f ./local-job2-2-verify.state 00:15:38.079 18:01:31 nvmf_tcp.nvmf_fio_target -- target/fio.sh@89 -- # trap - SIGINT SIGTERM EXIT 00:15:38.079 18:01:31 nvmf_tcp.nvmf_fio_target -- target/fio.sh@91 -- # nvmftestfini 00:15:38.079 18:01:31 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@488 -- # nvmfcleanup 00:15:38.079 18:01:31 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@117 -- # sync 00:15:38.079 18:01:31 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:15:38.079 18:01:31 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@120 -- # set +e 00:15:38.079 18:01:31 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@121 -- # for i in {1..20} 00:15:38.079 18:01:31 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:15:38.079 rmmod nvme_tcp 00:15:38.079 rmmod nvme_fabrics 00:15:38.079 rmmod nvme_keyring 00:15:38.338 18:01:31 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:15:38.338 18:01:31 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@124 -- # set -e 00:15:38.338 18:01:31 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@125 -- # return 0 00:15:38.338 18:01:31 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@489 -- # '[' -n 572769 ']' 00:15:38.338 18:01:31 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@490 -- # killprocess 572769 00:15:38.338 18:01:31 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@948 -- # '[' -z 572769 ']' 00:15:38.338 18:01:31 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@952 -- # kill -0 572769 00:15:38.338 18:01:31 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@953 -- # uname 00:15:38.338 18:01:31 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:15:38.338 18:01:31 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 572769 00:15:38.338 18:01:31 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:15:38.338 18:01:31 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:15:38.338 18:01:31 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@966 -- # echo 'killing process with pid 572769' 00:15:38.338 killing process with pid 572769 00:15:38.338 18:01:31 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@967 -- # kill 572769 00:15:38.338 18:01:31 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@972 -- # wait 572769 00:15:38.338 18:01:32 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:15:38.338 18:01:32 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:15:38.338 18:01:32 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:15:38.338 18:01:32 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:15:38.338 18:01:32 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@278 -- # remove_spdk_ns 00:15:38.338 18:01:32 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:38.338 18:01:32 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:15:38.338 18:01:32 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:40.862 18:01:34 nvmf_tcp.nvmf_fio_target -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:15:40.862 00:15:40.862 real 0m26.226s 00:15:40.862 user 1m46.632s 00:15:40.862 sys 0m7.885s 00:15:40.862 18:01:34 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@1124 -- # xtrace_disable 00:15:40.862 18:01:34 nvmf_tcp.nvmf_fio_target -- common/autotest_common.sh@10 -- # set +x 00:15:40.862 ************************************ 00:15:40.862 END TEST nvmf_fio_target 00:15:40.862 ************************************ 00:15:40.862 18:01:34 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:15:40.862 18:01:34 nvmf_tcp -- nvmf/nvmf.sh@56 -- # run_test nvmf_bdevio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp 00:15:40.862 18:01:34 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:15:40.862 18:01:34 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:15:40.862 18:01:34 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:15:40.862 ************************************ 00:15:40.862 START TEST nvmf_bdevio 00:15:40.862 ************************************ 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp 00:15:40.862 * Looking for test storage... 00:15:40.862 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- target/bdevio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@7 -- # uname -s 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- paths/export.sh@5 -- # export PATH 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@47 -- # : 0 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@51 -- # have_pci_nics=0 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- target/bdevio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:15:40.862 18:01:34 nvmf_tcp.nvmf_bdevio -- target/bdevio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:15:40.863 18:01:34 nvmf_tcp.nvmf_bdevio -- target/bdevio.sh@14 -- # nvmftestinit 00:15:40.863 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:15:40.863 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:15:40.863 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@448 -- # prepare_net_devs 00:15:40.863 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@410 -- # local -g is_hw=no 00:15:40.863 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@412 -- # remove_spdk_ns 00:15:40.863 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:40.863 18:01:34 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:15:40.863 18:01:34 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:40.863 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:15:40.863 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:15:40.863 18:01:34 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@285 -- # xtrace_disable 00:15:40.863 18:01:34 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@291 -- # pci_devs=() 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@291 -- # local -a pci_devs 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@292 -- # pci_net_devs=() 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@293 -- # pci_drivers=() 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@293 -- # local -A pci_drivers 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@295 -- # net_devs=() 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@295 -- # local -ga net_devs 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@296 -- # e810=() 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@296 -- # local -ga e810 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@297 -- # x722=() 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@297 -- # local -ga x722 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@298 -- # mlx=() 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@298 -- # local -ga mlx 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:15:46.116 Found 0000:86:00.0 (0x8086 - 0x159b) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:15:46.116 Found 0000:86:00.1 (0x8086 - 0x159b) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@390 -- # [[ up == up ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:15:46.116 Found net devices under 0000:86:00.0: cvl_0_0 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@390 -- # [[ up == up ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:15:46.116 Found net devices under 0000:86:00.1: cvl_0_1 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@414 -- # is_hw=yes 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:15:46.116 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:15:46.117 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:15:46.117 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.167 ms 00:15:46.117 00:15:46.117 --- 10.0.0.2 ping statistics --- 00:15:46.117 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:46.117 rtt min/avg/max/mdev = 0.167/0.167/0.167/0.000 ms 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:15:46.117 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:15:46.117 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.211 ms 00:15:46.117 00:15:46.117 --- 10.0.0.1 ping statistics --- 00:15:46.117 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:46.117 rtt min/avg/max/mdev = 0.211/0.211/0.211/0.000 ms 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@422 -- # return 0 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- target/bdevio.sh@16 -- # nvmfappstart -m 0x78 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@722 -- # xtrace_disable 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@481 -- # nvmfpid=579930 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x78 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@482 -- # waitforlisten 579930 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@829 -- # '[' -z 579930 ']' 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@834 -- # local max_retries=100 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:46.117 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@838 -- # xtrace_disable 00:15:46.117 18:01:39 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:15:46.117 [2024-07-15 18:01:39.459208] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:15:46.117 [2024-07-15 18:01:39.459258] [ DPDK EAL parameters: nvmf -c 0x78 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:15:46.117 EAL: No free 2048 kB hugepages reported on node 1 00:15:46.117 [2024-07-15 18:01:39.517261] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:15:46.117 [2024-07-15 18:01:39.589936] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:15:46.117 [2024-07-15 18:01:39.589976] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:15:46.117 [2024-07-15 18:01:39.589983] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:15:46.117 [2024-07-15 18:01:39.589988] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:15:46.117 [2024-07-15 18:01:39.589993] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:15:46.117 [2024-07-15 18:01:39.590132] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 4 00:15:46.117 [2024-07-15 18:01:39.590258] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 5 00:15:46.117 [2024-07-15 18:01:39.590367] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:15:46.117 [2024-07-15 18:01:39.590368] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 6 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@862 -- # return 0 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@728 -- # xtrace_disable 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- target/bdevio.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:15:46.683 [2024-07-15 18:01:40.309239] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- target/bdevio.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:15:46.683 Malloc0 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- target/bdevio.sh@20 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- target/bdevio.sh@21 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- target/bdevio.sh@22 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:15:46.683 [2024-07-15 18:01:40.360547] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- target/bdevio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/bdevio/bdevio --json /dev/fd/62 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- target/bdevio.sh@24 -- # gen_nvmf_target_json 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@532 -- # config=() 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@532 -- # local subsystem config 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:15:46.683 { 00:15:46.683 "params": { 00:15:46.683 "name": "Nvme$subsystem", 00:15:46.683 "trtype": "$TEST_TRANSPORT", 00:15:46.683 "traddr": "$NVMF_FIRST_TARGET_IP", 00:15:46.683 "adrfam": "ipv4", 00:15:46.683 "trsvcid": "$NVMF_PORT", 00:15:46.683 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:15:46.683 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:15:46.683 "hdgst": ${hdgst:-false}, 00:15:46.683 "ddgst": ${ddgst:-false} 00:15:46.683 }, 00:15:46.683 "method": "bdev_nvme_attach_controller" 00:15:46.683 } 00:15:46.683 EOF 00:15:46.683 )") 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@554 -- # cat 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@556 -- # jq . 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@557 -- # IFS=, 00:15:46.683 18:01:40 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:15:46.683 "params": { 00:15:46.683 "name": "Nvme1", 00:15:46.683 "trtype": "tcp", 00:15:46.683 "traddr": "10.0.0.2", 00:15:46.683 "adrfam": "ipv4", 00:15:46.683 "trsvcid": "4420", 00:15:46.683 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:15:46.683 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:15:46.683 "hdgst": false, 00:15:46.683 "ddgst": false 00:15:46.683 }, 00:15:46.683 "method": "bdev_nvme_attach_controller" 00:15:46.683 }' 00:15:46.683 [2024-07-15 18:01:40.407638] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:15:46.683 [2024-07-15 18:01:40.407683] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid580101 ] 00:15:46.942 EAL: No free 2048 kB hugepages reported on node 1 00:15:46.942 [2024-07-15 18:01:40.464388] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:15:46.942 [2024-07-15 18:01:40.542220] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:15:46.942 [2024-07-15 18:01:40.542316] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:15:46.942 [2024-07-15 18:01:40.542413] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:15:47.199 I/O targets: 00:15:47.199 Nvme1n1: 131072 blocks of 512 bytes (64 MiB) 00:15:47.199 00:15:47.199 00:15:47.199 CUnit - A unit testing framework for C - Version 2.1-3 00:15:47.199 http://cunit.sourceforge.net/ 00:15:47.199 00:15:47.199 00:15:47.199 Suite: bdevio tests on: Nvme1n1 00:15:47.199 Test: blockdev write read block ...passed 00:15:47.199 Test: blockdev write zeroes read block ...passed 00:15:47.199 Test: blockdev write zeroes read no split ...passed 00:15:47.199 Test: blockdev write zeroes read split ...passed 00:15:47.199 Test: blockdev write zeroes read split partial ...passed 00:15:47.199 Test: blockdev reset ...[2024-07-15 18:01:40.857915] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:15:47.199 [2024-07-15 18:01:40.857979] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x22dc6d0 (9): Bad file descriptor 00:15:47.199 [2024-07-15 18:01:40.918384] bdev_nvme.c:2067:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:15:47.199 passed 00:15:47.199 Test: blockdev write read 8 blocks ...passed 00:15:47.199 Test: blockdev write read size > 128k ...passed 00:15:47.199 Test: blockdev write read invalid size ...passed 00:15:47.457 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:15:47.457 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:15:47.457 Test: blockdev write read max offset ...passed 00:15:47.457 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:15:47.457 Test: blockdev writev readv 8 blocks ...passed 00:15:47.457 Test: blockdev writev readv 30 x 1block ...passed 00:15:47.457 Test: blockdev writev readv block ...passed 00:15:47.457 Test: blockdev writev readv size > 128k ...passed 00:15:47.457 Test: blockdev writev readv size > 128k in two iovs ...passed 00:15:47.457 Test: blockdev comparev and writev ...[2024-07-15 18:01:41.172360] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:15:47.457 [2024-07-15 18:01:41.172387] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:15:47.457 [2024-07-15 18:01:41.172400] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:15:47.457 [2024-07-15 18:01:41.172409] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:15:47.457 [2024-07-15 18:01:41.172677] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:15:47.457 [2024-07-15 18:01:41.172687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:15:47.457 [2024-07-15 18:01:41.172699] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:15:47.457 [2024-07-15 18:01:41.172707] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:15:47.457 [2024-07-15 18:01:41.172969] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:15:47.457 [2024-07-15 18:01:41.172978] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:15:47.457 [2024-07-15 18:01:41.172989] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:15:47.457 [2024-07-15 18:01:41.172996] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:15:47.457 [2024-07-15 18:01:41.173279] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:15:47.457 [2024-07-15 18:01:41.173290] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:15:47.457 [2024-07-15 18:01:41.173306] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:15:47.457 [2024-07-15 18:01:41.173313] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:15:47.715 passed 00:15:47.715 Test: blockdev nvme passthru rw ...passed 00:15:47.715 Test: blockdev nvme passthru vendor specific ...[2024-07-15 18:01:41.255723] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:15:47.715 [2024-07-15 18:01:41.255745] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:15:47.715 [2024-07-15 18:01:41.255888] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:15:47.715 [2024-07-15 18:01:41.255898] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:15:47.715 [2024-07-15 18:01:41.256032] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:15:47.715 [2024-07-15 18:01:41.256041] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:15:47.715 [2024-07-15 18:01:41.256173] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:15:47.715 [2024-07-15 18:01:41.256181] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:15:47.715 passed 00:15:47.715 Test: blockdev nvme admin passthru ...passed 00:15:47.715 Test: blockdev copy ...passed 00:15:47.715 00:15:47.715 Run Summary: Type Total Ran Passed Failed Inactive 00:15:47.715 suites 1 1 n/a 0 0 00:15:47.715 tests 23 23 23 0 0 00:15:47.715 asserts 152 152 152 0 n/a 00:15:47.715 00:15:47.715 Elapsed time = 1.223 seconds 00:15:47.972 18:01:41 nvmf_tcp.nvmf_bdevio -- target/bdevio.sh@26 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:15:47.972 18:01:41 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:47.972 18:01:41 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:15:47.972 18:01:41 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:47.972 18:01:41 nvmf_tcp.nvmf_bdevio -- target/bdevio.sh@28 -- # trap - SIGINT SIGTERM EXIT 00:15:47.972 18:01:41 nvmf_tcp.nvmf_bdevio -- target/bdevio.sh@30 -- # nvmftestfini 00:15:47.972 18:01:41 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@488 -- # nvmfcleanup 00:15:47.972 18:01:41 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@117 -- # sync 00:15:47.972 18:01:41 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:15:47.973 18:01:41 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@120 -- # set +e 00:15:47.973 18:01:41 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@121 -- # for i in {1..20} 00:15:47.973 18:01:41 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:15:47.973 rmmod nvme_tcp 00:15:47.973 rmmod nvme_fabrics 00:15:47.973 rmmod nvme_keyring 00:15:47.973 18:01:41 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:15:47.973 18:01:41 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@124 -- # set -e 00:15:47.973 18:01:41 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@125 -- # return 0 00:15:47.973 18:01:41 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@489 -- # '[' -n 579930 ']' 00:15:47.973 18:01:41 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@490 -- # killprocess 579930 00:15:47.973 18:01:41 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@948 -- # '[' -z 579930 ']' 00:15:47.973 18:01:41 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@952 -- # kill -0 579930 00:15:47.973 18:01:41 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@953 -- # uname 00:15:47.973 18:01:41 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:15:47.973 18:01:41 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 579930 00:15:47.973 18:01:41 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@954 -- # process_name=reactor_3 00:15:47.973 18:01:41 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@958 -- # '[' reactor_3 = sudo ']' 00:15:47.973 18:01:41 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@966 -- # echo 'killing process with pid 579930' 00:15:47.973 killing process with pid 579930 00:15:47.973 18:01:41 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@967 -- # kill 579930 00:15:47.973 18:01:41 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@972 -- # wait 579930 00:15:48.231 18:01:41 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:15:48.231 18:01:41 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:15:48.231 18:01:41 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:15:48.231 18:01:41 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:15:48.231 18:01:41 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@278 -- # remove_spdk_ns 00:15:48.231 18:01:41 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:48.231 18:01:41 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:15:48.231 18:01:41 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:50.164 18:01:43 nvmf_tcp.nvmf_bdevio -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:15:50.164 00:15:50.164 real 0m9.681s 00:15:50.164 user 0m12.265s 00:15:50.164 sys 0m4.356s 00:15:50.164 18:01:43 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@1124 -- # xtrace_disable 00:15:50.164 18:01:43 nvmf_tcp.nvmf_bdevio -- common/autotest_common.sh@10 -- # set +x 00:15:50.164 ************************************ 00:15:50.164 END TEST nvmf_bdevio 00:15:50.164 ************************************ 00:15:50.423 18:01:43 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:15:50.423 18:01:43 nvmf_tcp -- nvmf/nvmf.sh@57 -- # run_test nvmf_auth_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/auth.sh --transport=tcp 00:15:50.423 18:01:43 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:15:50.423 18:01:43 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:15:50.423 18:01:43 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:15:50.423 ************************************ 00:15:50.423 START TEST nvmf_auth_target 00:15:50.423 ************************************ 00:15:50.423 18:01:43 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/auth.sh --transport=tcp 00:15:50.423 * Looking for test storage... 00:15:50.423 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@7 -- # uname -s 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- paths/export.sh@5 -- # export PATH 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@47 -- # : 0 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@51 -- # have_pci_nics=0 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@13 -- # digests=("sha256" "sha384" "sha512") 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@14 -- # dhgroups=("null" "ffdhe2048" "ffdhe3072" "ffdhe4096" "ffdhe6144" "ffdhe8192") 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@15 -- # subnqn=nqn.2024-03.io.spdk:cnode0 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@16 -- # hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@17 -- # hostsock=/var/tmp/host.sock 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@18 -- # keys=() 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@18 -- # ckeys=() 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@59 -- # nvmftestinit 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@448 -- # prepare_net_devs 00:15:50.423 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@410 -- # local -g is_hw=no 00:15:50.424 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@412 -- # remove_spdk_ns 00:15:50.424 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:15:50.424 18:01:44 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:15:50.424 18:01:44 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:15:50.424 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:15:50.424 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:15:50.424 18:01:44 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@285 -- # xtrace_disable 00:15:50.424 18:01:44 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@291 -- # pci_devs=() 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@291 -- # local -a pci_devs 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@292 -- # pci_net_devs=() 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@293 -- # pci_drivers=() 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@293 -- # local -A pci_drivers 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@295 -- # net_devs=() 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@295 -- # local -ga net_devs 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@296 -- # e810=() 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@296 -- # local -ga e810 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@297 -- # x722=() 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@297 -- # local -ga x722 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@298 -- # mlx=() 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@298 -- # local -ga mlx 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:15:55.695 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:15:55.696 Found 0000:86:00.0 (0x8086 - 0x159b) 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:15:55.696 Found 0000:86:00.1 (0x8086 - 0x159b) 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@390 -- # [[ up == up ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:15:55.696 Found net devices under 0000:86:00.0: cvl_0_0 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@390 -- # [[ up == up ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:15:55.696 Found net devices under 0000:86:00.1: cvl_0_1 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@414 -- # is_hw=yes 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:15:55.696 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:15:55.696 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.211 ms 00:15:55.696 00:15:55.696 --- 10.0.0.2 ping statistics --- 00:15:55.696 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:55.696 rtt min/avg/max/mdev = 0.211/0.211/0.211/0.000 ms 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:15:55.696 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:15:55.696 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.233 ms 00:15:55.696 00:15:55.696 --- 10.0.0.1 ping statistics --- 00:15:55.696 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:15:55.696 rtt min/avg/max/mdev = 0.233/0.233/0.233/0.000 ms 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@422 -- # return 0 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@60 -- # nvmfappstart -L nvmf_auth 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@722 -- # xtrace_disable 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@481 -- # nvmfpid=583809 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@482 -- # waitforlisten 583809 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -L nvmf_auth 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@829 -- # '[' -z 583809 ']' 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@834 -- # local max_retries=100 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@838 -- # xtrace_disable 00:15:55.696 18:01:49 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@862 -- # return 0 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@728 -- # xtrace_disable 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt -m 2 -r /var/tmp/host.sock -L nvme_auth 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@62 -- # hostpid=583868 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@64 -- # trap 'dumplogs; cleanup' SIGINT SIGTERM EXIT 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@67 -- # gen_dhchap_key null 48 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@723 -- # local digest len file key 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@724 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@724 -- # local -A digests 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@726 -- # digest=null 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@726 -- # len=48 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@727 -- # xxd -p -c0 -l 24 /dev/urandom 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@727 -- # key=d745421a69a773238ae1543bc642de4bacd569f5954ae7ee 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@728 -- # mktemp -t spdk.key-null.XXX 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@728 -- # file=/tmp/spdk.key-null.mBL 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@729 -- # format_dhchap_key d745421a69a773238ae1543bc642de4bacd569f5954ae7ee 0 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@719 -- # format_key DHHC-1 d745421a69a773238ae1543bc642de4bacd569f5954ae7ee 0 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@702 -- # local prefix key digest 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # prefix=DHHC-1 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # key=d745421a69a773238ae1543bc642de4bacd569f5954ae7ee 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # digest=0 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@705 -- # python - 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@730 -- # chmod 0600 /tmp/spdk.key-null.mBL 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@732 -- # echo /tmp/spdk.key-null.mBL 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@67 -- # keys[0]=/tmp/spdk.key-null.mBL 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@67 -- # gen_dhchap_key sha512 64 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@723 -- # local digest len file key 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@724 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@724 -- # local -A digests 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@726 -- # digest=sha512 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@726 -- # len=64 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@727 -- # xxd -p -c0 -l 32 /dev/urandom 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@727 -- # key=6c48111f342e905f8259be8afd64f5a30dafbed1672ab64dd7be006dad61bb57 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@728 -- # mktemp -t spdk.key-sha512.XXX 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@728 -- # file=/tmp/spdk.key-sha512.AEk 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@729 -- # format_dhchap_key 6c48111f342e905f8259be8afd64f5a30dafbed1672ab64dd7be006dad61bb57 3 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@719 -- # format_key DHHC-1 6c48111f342e905f8259be8afd64f5a30dafbed1672ab64dd7be006dad61bb57 3 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@702 -- # local prefix key digest 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # prefix=DHHC-1 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # key=6c48111f342e905f8259be8afd64f5a30dafbed1672ab64dd7be006dad61bb57 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # digest=3 00:15:56.635 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@705 -- # python - 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@730 -- # chmod 0600 /tmp/spdk.key-sha512.AEk 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@732 -- # echo /tmp/spdk.key-sha512.AEk 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@67 -- # ckeys[0]=/tmp/spdk.key-sha512.AEk 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@68 -- # gen_dhchap_key sha256 32 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@723 -- # local digest len file key 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@724 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@724 -- # local -A digests 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@726 -- # digest=sha256 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@726 -- # len=32 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@727 -- # xxd -p -c0 -l 16 /dev/urandom 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@727 -- # key=d3366a819d90f75c2e1da32f73defc3d 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@728 -- # mktemp -t spdk.key-sha256.XXX 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@728 -- # file=/tmp/spdk.key-sha256.aI0 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@729 -- # format_dhchap_key d3366a819d90f75c2e1da32f73defc3d 1 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@719 -- # format_key DHHC-1 d3366a819d90f75c2e1da32f73defc3d 1 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@702 -- # local prefix key digest 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # prefix=DHHC-1 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # key=d3366a819d90f75c2e1da32f73defc3d 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # digest=1 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@705 -- # python - 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@730 -- # chmod 0600 /tmp/spdk.key-sha256.aI0 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@732 -- # echo /tmp/spdk.key-sha256.aI0 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@68 -- # keys[1]=/tmp/spdk.key-sha256.aI0 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@68 -- # gen_dhchap_key sha384 48 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@723 -- # local digest len file key 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@724 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@724 -- # local -A digests 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@726 -- # digest=sha384 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@726 -- # len=48 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@727 -- # xxd -p -c0 -l 24 /dev/urandom 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@727 -- # key=9aff74098d2fec2aac6bf6c536d25006ecdf8c164176dd7e 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@728 -- # mktemp -t spdk.key-sha384.XXX 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@728 -- # file=/tmp/spdk.key-sha384.TJQ 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@729 -- # format_dhchap_key 9aff74098d2fec2aac6bf6c536d25006ecdf8c164176dd7e 2 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@719 -- # format_key DHHC-1 9aff74098d2fec2aac6bf6c536d25006ecdf8c164176dd7e 2 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@702 -- # local prefix key digest 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # prefix=DHHC-1 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # key=9aff74098d2fec2aac6bf6c536d25006ecdf8c164176dd7e 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # digest=2 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@705 -- # python - 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@730 -- # chmod 0600 /tmp/spdk.key-sha384.TJQ 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@732 -- # echo /tmp/spdk.key-sha384.TJQ 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@68 -- # ckeys[1]=/tmp/spdk.key-sha384.TJQ 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@69 -- # gen_dhchap_key sha384 48 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@723 -- # local digest len file key 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@724 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:15:56.895 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@724 -- # local -A digests 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@726 -- # digest=sha384 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@726 -- # len=48 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@727 -- # xxd -p -c0 -l 24 /dev/urandom 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@727 -- # key=3bcd365cafd914c7e0cd6b8b9e0ba536b73bd6066adbdb37 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@728 -- # mktemp -t spdk.key-sha384.XXX 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@728 -- # file=/tmp/spdk.key-sha384.pOl 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@729 -- # format_dhchap_key 3bcd365cafd914c7e0cd6b8b9e0ba536b73bd6066adbdb37 2 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@719 -- # format_key DHHC-1 3bcd365cafd914c7e0cd6b8b9e0ba536b73bd6066adbdb37 2 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@702 -- # local prefix key digest 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # prefix=DHHC-1 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # key=3bcd365cafd914c7e0cd6b8b9e0ba536b73bd6066adbdb37 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # digest=2 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@705 -- # python - 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@730 -- # chmod 0600 /tmp/spdk.key-sha384.pOl 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@732 -- # echo /tmp/spdk.key-sha384.pOl 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@69 -- # keys[2]=/tmp/spdk.key-sha384.pOl 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@69 -- # gen_dhchap_key sha256 32 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@723 -- # local digest len file key 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@724 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@724 -- # local -A digests 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@726 -- # digest=sha256 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@726 -- # len=32 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@727 -- # xxd -p -c0 -l 16 /dev/urandom 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@727 -- # key=91e9927c676befd5dc918fe1552b2784 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@728 -- # mktemp -t spdk.key-sha256.XXX 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@728 -- # file=/tmp/spdk.key-sha256.7IT 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@729 -- # format_dhchap_key 91e9927c676befd5dc918fe1552b2784 1 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@719 -- # format_key DHHC-1 91e9927c676befd5dc918fe1552b2784 1 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@702 -- # local prefix key digest 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # prefix=DHHC-1 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # key=91e9927c676befd5dc918fe1552b2784 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # digest=1 00:15:56.896 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@705 -- # python - 00:15:57.154 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@730 -- # chmod 0600 /tmp/spdk.key-sha256.7IT 00:15:57.154 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@732 -- # echo /tmp/spdk.key-sha256.7IT 00:15:57.154 18:01:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@69 -- # ckeys[2]=/tmp/spdk.key-sha256.7IT 00:15:57.154 18:01:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@70 -- # gen_dhchap_key sha512 64 00:15:57.154 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@723 -- # local digest len file key 00:15:57.154 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@724 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:15:57.154 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@724 -- # local -A digests 00:15:57.154 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@726 -- # digest=sha512 00:15:57.154 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@726 -- # len=64 00:15:57.154 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@727 -- # xxd -p -c0 -l 32 /dev/urandom 00:15:57.154 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@727 -- # key=092e957e7b5e02c58f900033cf3ed7f6ba96d07373a39182461269f0c09a4c3e 00:15:57.154 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@728 -- # mktemp -t spdk.key-sha512.XXX 00:15:57.154 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@728 -- # file=/tmp/spdk.key-sha512.IyM 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@729 -- # format_dhchap_key 092e957e7b5e02c58f900033cf3ed7f6ba96d07373a39182461269f0c09a4c3e 3 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@719 -- # format_key DHHC-1 092e957e7b5e02c58f900033cf3ed7f6ba96d07373a39182461269f0c09a4c3e 3 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@702 -- # local prefix key digest 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # prefix=DHHC-1 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # key=092e957e7b5e02c58f900033cf3ed7f6ba96d07373a39182461269f0c09a4c3e 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@704 -- # digest=3 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@705 -- # python - 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@730 -- # chmod 0600 /tmp/spdk.key-sha512.IyM 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@732 -- # echo /tmp/spdk.key-sha512.IyM 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@70 -- # keys[3]=/tmp/spdk.key-sha512.IyM 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@70 -- # ckeys[3]= 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@72 -- # waitforlisten 583809 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@829 -- # '[' -z 583809 ']' 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@834 -- # local max_retries=100 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:15:57.155 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@838 -- # xtrace_disable 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@862 -- # return 0 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@73 -- # waitforlisten 583868 /var/tmp/host.sock 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@829 -- # '[' -z 583868 ']' 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/host.sock 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@834 -- # local max_retries=100 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/host.sock...' 00:15:57.155 Waiting for process to start up and listen on UNIX domain socket /var/tmp/host.sock... 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@838 -- # xtrace_disable 00:15:57.155 18:01:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:57.413 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:15:57.413 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@862 -- # return 0 00:15:57.413 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@74 -- # rpc_cmd 00:15:57.413 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:57.413 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:57.413 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:57.413 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@81 -- # for i in "${!keys[@]}" 00:15:57.413 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@82 -- # rpc_cmd keyring_file_add_key key0 /tmp/spdk.key-null.mBL 00:15:57.413 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:57.413 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:57.413 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:57.413 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@83 -- # hostrpc keyring_file_add_key key0 /tmp/spdk.key-null.mBL 00:15:57.413 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key key0 /tmp/spdk.key-null.mBL 00:15:57.672 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@84 -- # [[ -n /tmp/spdk.key-sha512.AEk ]] 00:15:57.672 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@85 -- # rpc_cmd keyring_file_add_key ckey0 /tmp/spdk.key-sha512.AEk 00:15:57.673 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:57.673 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:57.673 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:57.673 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@86 -- # hostrpc keyring_file_add_key ckey0 /tmp/spdk.key-sha512.AEk 00:15:57.673 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key ckey0 /tmp/spdk.key-sha512.AEk 00:15:57.932 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@81 -- # for i in "${!keys[@]}" 00:15:57.932 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@82 -- # rpc_cmd keyring_file_add_key key1 /tmp/spdk.key-sha256.aI0 00:15:57.932 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:57.932 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:57.932 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:57.932 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@83 -- # hostrpc keyring_file_add_key key1 /tmp/spdk.key-sha256.aI0 00:15:57.932 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key key1 /tmp/spdk.key-sha256.aI0 00:15:57.932 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@84 -- # [[ -n /tmp/spdk.key-sha384.TJQ ]] 00:15:57.932 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@85 -- # rpc_cmd keyring_file_add_key ckey1 /tmp/spdk.key-sha384.TJQ 00:15:57.932 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:57.932 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:57.932 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:57.932 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@86 -- # hostrpc keyring_file_add_key ckey1 /tmp/spdk.key-sha384.TJQ 00:15:57.932 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key ckey1 /tmp/spdk.key-sha384.TJQ 00:15:58.190 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@81 -- # for i in "${!keys[@]}" 00:15:58.190 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@82 -- # rpc_cmd keyring_file_add_key key2 /tmp/spdk.key-sha384.pOl 00:15:58.190 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:58.190 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:58.190 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:58.190 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@83 -- # hostrpc keyring_file_add_key key2 /tmp/spdk.key-sha384.pOl 00:15:58.190 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key key2 /tmp/spdk.key-sha384.pOl 00:15:58.449 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@84 -- # [[ -n /tmp/spdk.key-sha256.7IT ]] 00:15:58.449 18:01:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@85 -- # rpc_cmd keyring_file_add_key ckey2 /tmp/spdk.key-sha256.7IT 00:15:58.449 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:58.449 18:01:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:58.449 18:01:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:58.449 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@86 -- # hostrpc keyring_file_add_key ckey2 /tmp/spdk.key-sha256.7IT 00:15:58.449 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key ckey2 /tmp/spdk.key-sha256.7IT 00:15:58.707 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@81 -- # for i in "${!keys[@]}" 00:15:58.707 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@82 -- # rpc_cmd keyring_file_add_key key3 /tmp/spdk.key-sha512.IyM 00:15:58.707 18:01:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:58.707 18:01:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:58.707 18:01:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:58.707 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@83 -- # hostrpc keyring_file_add_key key3 /tmp/spdk.key-sha512.IyM 00:15:58.707 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock keyring_file_add_key key3 /tmp/spdk.key-sha512.IyM 00:15:58.707 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@84 -- # [[ -n '' ]] 00:15:58.707 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@91 -- # for digest in "${digests[@]}" 00:15:58.707 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@92 -- # for dhgroup in "${dhgroups[@]}" 00:15:58.707 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:15:58.707 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:15:58.707 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:15:58.965 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 null 0 00:15:58.965 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:15:58.965 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:15:58.965 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=null 00:15:58.965 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key0 00:15:58.965 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:15:58.965 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:15:58.965 18:01:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:58.965 18:01:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:58.965 18:01:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:58.965 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:15:58.965 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:15:59.224 00:15:59.224 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:15:59.224 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:15:59.224 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:15:59.224 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:15:59.483 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:15:59.483 18:01:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:15:59.483 18:01:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:15:59.483 18:01:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:15:59.483 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:15:59.483 { 00:15:59.483 "cntlid": 1, 00:15:59.483 "qid": 0, 00:15:59.483 "state": "enabled", 00:15:59.483 "thread": "nvmf_tgt_poll_group_000", 00:15:59.483 "listen_address": { 00:15:59.483 "trtype": "TCP", 00:15:59.483 "adrfam": "IPv4", 00:15:59.483 "traddr": "10.0.0.2", 00:15:59.483 "trsvcid": "4420" 00:15:59.483 }, 00:15:59.483 "peer_address": { 00:15:59.483 "trtype": "TCP", 00:15:59.483 "adrfam": "IPv4", 00:15:59.483 "traddr": "10.0.0.1", 00:15:59.483 "trsvcid": "53574" 00:15:59.483 }, 00:15:59.483 "auth": { 00:15:59.483 "state": "completed", 00:15:59.483 "digest": "sha256", 00:15:59.483 "dhgroup": "null" 00:15:59.483 } 00:15:59.483 } 00:15:59.483 ]' 00:15:59.483 18:01:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:15:59.483 18:01:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:15:59.483 18:01:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:15:59.483 18:01:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ null == \n\u\l\l ]] 00:15:59.483 18:01:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:15:59.483 18:01:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:15:59.483 18:01:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:15:59.483 18:01:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:15:59.742 18:01:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:00:ZDc0NTQyMWE2OWE3NzMyMzhhZTE1NDNiYzY0MmRlNGJhY2Q1NjlmNTk1NGFlN2VlXS98/Q==: --dhchap-ctrl-secret DHHC-1:03:NmM0ODExMWYzNDJlOTA1ZjgyNTliZThhZmQ2NGY1YTMwZGFmYmVkMTY3MmFiNjRkZDdiZTAwNmRhZDYxYmI1N4Bp9x8=: 00:16:00.308 18:01:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:00.308 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:00.308 18:01:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:00.308 18:01:53 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:00.308 18:01:53 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:00.308 18:01:53 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:00.308 18:01:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:00.308 18:01:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:16:00.308 18:01:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:16:00.566 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 null 1 00:16:00.566 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:00.566 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:00.566 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=null 00:16:00.566 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key1 00:16:00.566 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:00.566 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:00.566 18:01:54 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:00.566 18:01:54 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:00.566 18:01:54 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:00.566 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:00.566 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:00.566 00:16:00.566 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:00.566 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:00.566 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:00.825 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:00.825 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:00.825 18:01:54 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:00.825 18:01:54 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:00.825 18:01:54 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:00.825 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:00.825 { 00:16:00.825 "cntlid": 3, 00:16:00.826 "qid": 0, 00:16:00.826 "state": "enabled", 00:16:00.826 "thread": "nvmf_tgt_poll_group_000", 00:16:00.826 "listen_address": { 00:16:00.826 "trtype": "TCP", 00:16:00.826 "adrfam": "IPv4", 00:16:00.826 "traddr": "10.0.0.2", 00:16:00.826 "trsvcid": "4420" 00:16:00.826 }, 00:16:00.826 "peer_address": { 00:16:00.826 "trtype": "TCP", 00:16:00.826 "adrfam": "IPv4", 00:16:00.826 "traddr": "10.0.0.1", 00:16:00.826 "trsvcid": "53608" 00:16:00.826 }, 00:16:00.826 "auth": { 00:16:00.826 "state": "completed", 00:16:00.826 "digest": "sha256", 00:16:00.826 "dhgroup": "null" 00:16:00.826 } 00:16:00.826 } 00:16:00.826 ]' 00:16:00.826 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:00.826 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:00.826 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:00.826 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ null == \n\u\l\l ]] 00:16:01.085 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:01.085 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:01.085 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:01.085 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:01.085 18:01:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:01:ZDMzNjZhODE5ZDkwZjc1YzJlMWRhMzJmNzNkZWZjM2TGg2/J: --dhchap-ctrl-secret DHHC-1:02:OWFmZjc0MDk4ZDJmZWMyYWFjNmJmNmM1MzZkMjUwMDZlY2RmOGMxNjQxNzZkZDdloY1p5Q==: 00:16:01.652 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:01.652 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:01.652 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:01.652 18:01:55 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:01.652 18:01:55 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:01.652 18:01:55 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:01.652 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:01.652 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:16:01.652 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:16:01.910 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 null 2 00:16:01.911 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:01.911 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:01.911 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=null 00:16:01.911 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key2 00:16:01.911 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:01.911 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:01.911 18:01:55 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:01.911 18:01:55 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:01.911 18:01:55 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:01.911 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:01.911 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:02.169 00:16:02.169 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:02.169 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:02.169 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:02.427 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:02.427 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:02.427 18:01:55 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:02.427 18:01:55 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:02.427 18:01:55 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:02.427 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:02.427 { 00:16:02.428 "cntlid": 5, 00:16:02.428 "qid": 0, 00:16:02.428 "state": "enabled", 00:16:02.428 "thread": "nvmf_tgt_poll_group_000", 00:16:02.428 "listen_address": { 00:16:02.428 "trtype": "TCP", 00:16:02.428 "adrfam": "IPv4", 00:16:02.428 "traddr": "10.0.0.2", 00:16:02.428 "trsvcid": "4420" 00:16:02.428 }, 00:16:02.428 "peer_address": { 00:16:02.428 "trtype": "TCP", 00:16:02.428 "adrfam": "IPv4", 00:16:02.428 "traddr": "10.0.0.1", 00:16:02.428 "trsvcid": "53634" 00:16:02.428 }, 00:16:02.428 "auth": { 00:16:02.428 "state": "completed", 00:16:02.428 "digest": "sha256", 00:16:02.428 "dhgroup": "null" 00:16:02.428 } 00:16:02.428 } 00:16:02.428 ]' 00:16:02.428 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:02.428 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:02.428 18:01:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:02.428 18:01:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ null == \n\u\l\l ]] 00:16:02.428 18:01:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:02.428 18:01:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:02.428 18:01:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:02.428 18:01:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:02.687 18:01:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:02:M2JjZDM2NWNhZmQ5MTRjN2UwY2Q2YjhiOWUwYmE1MzZiNzNiZDYwNjZhZGJkYjM3IIKM4Q==: --dhchap-ctrl-secret DHHC-1:01:OTFlOTkyN2M2NzZiZWZkNWRjOTE4ZmUxNTUyYjI3ODS55lrp: 00:16:03.254 18:01:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:03.254 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:03.254 18:01:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:03.254 18:01:56 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:03.254 18:01:56 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:03.254 18:01:56 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:03.254 18:01:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:03.254 18:01:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:16:03.254 18:01:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups null 00:16:03.514 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 null 3 00:16:03.514 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:03.514 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:03.514 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=null 00:16:03.514 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key3 00:16:03.514 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:03.514 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key3 00:16:03.514 18:01:57 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:03.514 18:01:57 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:03.514 18:01:57 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:03.514 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:16:03.514 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:16:03.773 00:16:03.773 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:03.773 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:03.773 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:03.773 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:03.773 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:03.773 18:01:57 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:03.773 18:01:57 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:03.773 18:01:57 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:03.773 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:03.773 { 00:16:03.773 "cntlid": 7, 00:16:03.773 "qid": 0, 00:16:03.773 "state": "enabled", 00:16:03.773 "thread": "nvmf_tgt_poll_group_000", 00:16:03.773 "listen_address": { 00:16:03.773 "trtype": "TCP", 00:16:03.773 "adrfam": "IPv4", 00:16:03.773 "traddr": "10.0.0.2", 00:16:03.773 "trsvcid": "4420" 00:16:03.773 }, 00:16:03.773 "peer_address": { 00:16:03.773 "trtype": "TCP", 00:16:03.773 "adrfam": "IPv4", 00:16:03.773 "traddr": "10.0.0.1", 00:16:03.773 "trsvcid": "48956" 00:16:03.773 }, 00:16:03.773 "auth": { 00:16:03.773 "state": "completed", 00:16:03.773 "digest": "sha256", 00:16:03.773 "dhgroup": "null" 00:16:03.773 } 00:16:03.773 } 00:16:03.773 ]' 00:16:03.773 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:03.773 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:04.031 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:04.031 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ null == \n\u\l\l ]] 00:16:04.031 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:04.031 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:04.031 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:04.031 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:04.031 18:01:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:03:MDkyZTk1N2U3YjVlMDJjNThmOTAwMDMzY2YzZWQ3ZjZiYTk2ZDA3MzczYTM5MTgyNDYxMjY5ZjBjMDlhNGMzZRaivds=: 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:04.967 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@92 -- # for dhgroup in "${dhgroups[@]}" 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 ffdhe2048 0 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe2048 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key0 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:04.967 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:05.226 00:16:05.226 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:05.226 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:05.226 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:05.486 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:05.486 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:05.486 18:01:58 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:05.486 18:01:58 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:05.486 18:01:58 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:05.486 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:05.486 { 00:16:05.486 "cntlid": 9, 00:16:05.486 "qid": 0, 00:16:05.486 "state": "enabled", 00:16:05.486 "thread": "nvmf_tgt_poll_group_000", 00:16:05.486 "listen_address": { 00:16:05.486 "trtype": "TCP", 00:16:05.486 "adrfam": "IPv4", 00:16:05.486 "traddr": "10.0.0.2", 00:16:05.486 "trsvcid": "4420" 00:16:05.486 }, 00:16:05.486 "peer_address": { 00:16:05.486 "trtype": "TCP", 00:16:05.486 "adrfam": "IPv4", 00:16:05.486 "traddr": "10.0.0.1", 00:16:05.486 "trsvcid": "49000" 00:16:05.486 }, 00:16:05.486 "auth": { 00:16:05.486 "state": "completed", 00:16:05.486 "digest": "sha256", 00:16:05.486 "dhgroup": "ffdhe2048" 00:16:05.486 } 00:16:05.486 } 00:16:05.486 ]' 00:16:05.486 18:01:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:05.486 18:01:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:05.486 18:01:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:05.486 18:01:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:16:05.486 18:01:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:05.486 18:01:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:05.486 18:01:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:05.486 18:01:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:05.744 18:01:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:00:ZDc0NTQyMWE2OWE3NzMyMzhhZTE1NDNiYzY0MmRlNGJhY2Q1NjlmNTk1NGFlN2VlXS98/Q==: --dhchap-ctrl-secret DHHC-1:03:NmM0ODExMWYzNDJlOTA1ZjgyNTliZThhZmQ2NGY1YTMwZGFmYmVkMTY3MmFiNjRkZDdiZTAwNmRhZDYxYmI1N4Bp9x8=: 00:16:06.310 18:01:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:06.310 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:06.310 18:01:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:06.310 18:01:59 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:06.310 18:01:59 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:06.310 18:01:59 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:06.310 18:01:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:06.310 18:01:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:16:06.310 18:01:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:16:06.310 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 ffdhe2048 1 00:16:06.310 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:06.310 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:06.310 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe2048 00:16:06.310 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key1 00:16:06.310 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:06.310 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:06.310 18:02:00 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:06.310 18:02:00 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:06.310 18:02:00 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:06.310 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:06.310 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:06.585 00:16:06.585 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:06.585 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:06.585 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:06.861 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:06.861 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:06.861 18:02:00 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:06.861 18:02:00 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:06.861 18:02:00 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:06.861 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:06.861 { 00:16:06.861 "cntlid": 11, 00:16:06.861 "qid": 0, 00:16:06.861 "state": "enabled", 00:16:06.861 "thread": "nvmf_tgt_poll_group_000", 00:16:06.861 "listen_address": { 00:16:06.861 "trtype": "TCP", 00:16:06.861 "adrfam": "IPv4", 00:16:06.861 "traddr": "10.0.0.2", 00:16:06.861 "trsvcid": "4420" 00:16:06.861 }, 00:16:06.861 "peer_address": { 00:16:06.861 "trtype": "TCP", 00:16:06.861 "adrfam": "IPv4", 00:16:06.861 "traddr": "10.0.0.1", 00:16:06.861 "trsvcid": "49022" 00:16:06.861 }, 00:16:06.861 "auth": { 00:16:06.861 "state": "completed", 00:16:06.861 "digest": "sha256", 00:16:06.861 "dhgroup": "ffdhe2048" 00:16:06.861 } 00:16:06.861 } 00:16:06.861 ]' 00:16:06.861 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:06.861 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:06.861 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:06.861 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:16:06.861 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:07.119 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:07.119 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:07.119 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:07.119 18:02:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:01:ZDMzNjZhODE5ZDkwZjc1YzJlMWRhMzJmNzNkZWZjM2TGg2/J: --dhchap-ctrl-secret DHHC-1:02:OWFmZjc0MDk4ZDJmZWMyYWFjNmJmNmM1MzZkMjUwMDZlY2RmOGMxNjQxNzZkZDdloY1p5Q==: 00:16:07.686 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:07.686 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:07.686 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:07.686 18:02:01 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:07.686 18:02:01 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:07.686 18:02:01 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:07.686 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:07.686 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:16:07.686 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:16:07.944 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 ffdhe2048 2 00:16:07.945 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:07.945 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:07.945 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe2048 00:16:07.945 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key2 00:16:07.945 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:07.945 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:07.945 18:02:01 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:07.945 18:02:01 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:07.945 18:02:01 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:07.945 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:07.945 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:08.203 00:16:08.203 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:08.203 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:08.203 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:08.462 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:08.462 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:08.462 18:02:01 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:08.462 18:02:01 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:08.462 18:02:01 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:08.462 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:08.462 { 00:16:08.462 "cntlid": 13, 00:16:08.462 "qid": 0, 00:16:08.462 "state": "enabled", 00:16:08.462 "thread": "nvmf_tgt_poll_group_000", 00:16:08.462 "listen_address": { 00:16:08.462 "trtype": "TCP", 00:16:08.462 "adrfam": "IPv4", 00:16:08.462 "traddr": "10.0.0.2", 00:16:08.462 "trsvcid": "4420" 00:16:08.462 }, 00:16:08.462 "peer_address": { 00:16:08.462 "trtype": "TCP", 00:16:08.462 "adrfam": "IPv4", 00:16:08.462 "traddr": "10.0.0.1", 00:16:08.462 "trsvcid": "49042" 00:16:08.462 }, 00:16:08.462 "auth": { 00:16:08.462 "state": "completed", 00:16:08.462 "digest": "sha256", 00:16:08.462 "dhgroup": "ffdhe2048" 00:16:08.462 } 00:16:08.462 } 00:16:08.462 ]' 00:16:08.462 18:02:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:08.462 18:02:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:08.462 18:02:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:08.462 18:02:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:16:08.462 18:02:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:08.462 18:02:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:08.462 18:02:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:08.462 18:02:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:08.749 18:02:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:02:M2JjZDM2NWNhZmQ5MTRjN2UwY2Q2YjhiOWUwYmE1MzZiNzNiZDYwNjZhZGJkYjM3IIKM4Q==: --dhchap-ctrl-secret DHHC-1:01:OTFlOTkyN2M2NzZiZWZkNWRjOTE4ZmUxNTUyYjI3ODS55lrp: 00:16:09.315 18:02:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:09.315 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:09.315 18:02:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:09.315 18:02:02 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:09.315 18:02:02 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:09.315 18:02:02 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:09.315 18:02:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:09.315 18:02:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:16:09.315 18:02:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:16:09.315 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 ffdhe2048 3 00:16:09.315 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:09.315 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:09.315 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe2048 00:16:09.315 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key3 00:16:09.315 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:09.315 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key3 00:16:09.315 18:02:03 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:09.315 18:02:03 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:09.315 18:02:03 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:09.315 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:16:09.315 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:16:09.573 00:16:09.573 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:09.573 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:09.573 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:09.830 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:09.830 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:09.830 18:02:03 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:09.830 18:02:03 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:09.830 18:02:03 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:09.830 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:09.830 { 00:16:09.830 "cntlid": 15, 00:16:09.830 "qid": 0, 00:16:09.830 "state": "enabled", 00:16:09.830 "thread": "nvmf_tgt_poll_group_000", 00:16:09.830 "listen_address": { 00:16:09.830 "trtype": "TCP", 00:16:09.830 "adrfam": "IPv4", 00:16:09.830 "traddr": "10.0.0.2", 00:16:09.830 "trsvcid": "4420" 00:16:09.830 }, 00:16:09.830 "peer_address": { 00:16:09.830 "trtype": "TCP", 00:16:09.830 "adrfam": "IPv4", 00:16:09.830 "traddr": "10.0.0.1", 00:16:09.830 "trsvcid": "49058" 00:16:09.830 }, 00:16:09.830 "auth": { 00:16:09.830 "state": "completed", 00:16:09.830 "digest": "sha256", 00:16:09.830 "dhgroup": "ffdhe2048" 00:16:09.830 } 00:16:09.830 } 00:16:09.830 ]' 00:16:09.830 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:09.830 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:09.830 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:09.830 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:16:09.830 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:10.088 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:10.088 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:10.088 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:10.088 18:02:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:03:MDkyZTk1N2U3YjVlMDJjNThmOTAwMDMzY2YzZWQ3ZjZiYTk2ZDA3MzczYTM5MTgyNDYxMjY5ZjBjMDlhNGMzZRaivds=: 00:16:10.654 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:10.654 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:10.654 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:10.654 18:02:04 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:10.654 18:02:04 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:10.654 18:02:04 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:10.654 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@92 -- # for dhgroup in "${dhgroups[@]}" 00:16:10.654 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:10.654 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:16:10.654 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:16:10.912 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 ffdhe3072 0 00:16:10.912 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:10.912 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:10.912 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe3072 00:16:10.912 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key0 00:16:10.912 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:10.912 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:10.912 18:02:04 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:10.912 18:02:04 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:10.912 18:02:04 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:10.912 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:10.912 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:11.170 00:16:11.170 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:11.170 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:11.170 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:11.427 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:11.427 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:11.427 18:02:04 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:11.427 18:02:04 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:11.427 18:02:04 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:11.427 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:11.427 { 00:16:11.427 "cntlid": 17, 00:16:11.427 "qid": 0, 00:16:11.427 "state": "enabled", 00:16:11.427 "thread": "nvmf_tgt_poll_group_000", 00:16:11.427 "listen_address": { 00:16:11.427 "trtype": "TCP", 00:16:11.427 "adrfam": "IPv4", 00:16:11.427 "traddr": "10.0.0.2", 00:16:11.427 "trsvcid": "4420" 00:16:11.427 }, 00:16:11.427 "peer_address": { 00:16:11.427 "trtype": "TCP", 00:16:11.427 "adrfam": "IPv4", 00:16:11.427 "traddr": "10.0.0.1", 00:16:11.427 "trsvcid": "49094" 00:16:11.427 }, 00:16:11.427 "auth": { 00:16:11.427 "state": "completed", 00:16:11.427 "digest": "sha256", 00:16:11.427 "dhgroup": "ffdhe3072" 00:16:11.427 } 00:16:11.427 } 00:16:11.427 ]' 00:16:11.427 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:11.427 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:11.427 18:02:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:11.427 18:02:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:16:11.427 18:02:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:11.427 18:02:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:11.427 18:02:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:11.427 18:02:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:11.684 18:02:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:00:ZDc0NTQyMWE2OWE3NzMyMzhhZTE1NDNiYzY0MmRlNGJhY2Q1NjlmNTk1NGFlN2VlXS98/Q==: --dhchap-ctrl-secret DHHC-1:03:NmM0ODExMWYzNDJlOTA1ZjgyNTliZThhZmQ2NGY1YTMwZGFmYmVkMTY3MmFiNjRkZDdiZTAwNmRhZDYxYmI1N4Bp9x8=: 00:16:12.249 18:02:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:12.249 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:12.249 18:02:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:12.249 18:02:05 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:12.249 18:02:05 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:12.249 18:02:05 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:12.249 18:02:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:12.249 18:02:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:16:12.249 18:02:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:16:12.249 18:02:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 ffdhe3072 1 00:16:12.249 18:02:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:12.249 18:02:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:12.249 18:02:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe3072 00:16:12.249 18:02:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key1 00:16:12.249 18:02:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:12.249 18:02:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:12.249 18:02:05 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:12.249 18:02:05 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:12.249 18:02:05 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:12.249 18:02:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:12.249 18:02:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:12.507 00:16:12.507 18:02:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:12.507 18:02:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:12.507 18:02:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:12.766 18:02:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:12.766 18:02:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:12.766 18:02:06 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:12.766 18:02:06 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:12.766 18:02:06 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:12.766 18:02:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:12.766 { 00:16:12.766 "cntlid": 19, 00:16:12.766 "qid": 0, 00:16:12.766 "state": "enabled", 00:16:12.766 "thread": "nvmf_tgt_poll_group_000", 00:16:12.766 "listen_address": { 00:16:12.766 "trtype": "TCP", 00:16:12.766 "adrfam": "IPv4", 00:16:12.766 "traddr": "10.0.0.2", 00:16:12.766 "trsvcid": "4420" 00:16:12.766 }, 00:16:12.766 "peer_address": { 00:16:12.766 "trtype": "TCP", 00:16:12.766 "adrfam": "IPv4", 00:16:12.766 "traddr": "10.0.0.1", 00:16:12.766 "trsvcid": "49124" 00:16:12.766 }, 00:16:12.766 "auth": { 00:16:12.766 "state": "completed", 00:16:12.766 "digest": "sha256", 00:16:12.766 "dhgroup": "ffdhe3072" 00:16:12.766 } 00:16:12.766 } 00:16:12.766 ]' 00:16:12.766 18:02:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:12.766 18:02:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:12.766 18:02:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:12.766 18:02:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:16:12.766 18:02:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:13.023 18:02:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:13.023 18:02:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:13.023 18:02:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:13.023 18:02:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:01:ZDMzNjZhODE5ZDkwZjc1YzJlMWRhMzJmNzNkZWZjM2TGg2/J: --dhchap-ctrl-secret DHHC-1:02:OWFmZjc0MDk4ZDJmZWMyYWFjNmJmNmM1MzZkMjUwMDZlY2RmOGMxNjQxNzZkZDdloY1p5Q==: 00:16:13.587 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:13.587 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:13.587 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:13.587 18:02:07 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:13.587 18:02:07 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:13.587 18:02:07 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:13.587 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:13.587 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:16:13.587 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:16:13.845 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 ffdhe3072 2 00:16:13.845 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:13.845 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:13.845 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe3072 00:16:13.845 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key2 00:16:13.845 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:13.845 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:13.845 18:02:07 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:13.845 18:02:07 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:13.845 18:02:07 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:13.845 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:13.845 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:14.103 00:16:14.103 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:14.103 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:14.103 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:14.360 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:14.360 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:14.360 18:02:07 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:14.360 18:02:07 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:14.360 18:02:07 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:14.360 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:14.360 { 00:16:14.360 "cntlid": 21, 00:16:14.360 "qid": 0, 00:16:14.360 "state": "enabled", 00:16:14.360 "thread": "nvmf_tgt_poll_group_000", 00:16:14.360 "listen_address": { 00:16:14.360 "trtype": "TCP", 00:16:14.360 "adrfam": "IPv4", 00:16:14.360 "traddr": "10.0.0.2", 00:16:14.360 "trsvcid": "4420" 00:16:14.360 }, 00:16:14.360 "peer_address": { 00:16:14.360 "trtype": "TCP", 00:16:14.360 "adrfam": "IPv4", 00:16:14.360 "traddr": "10.0.0.1", 00:16:14.360 "trsvcid": "41068" 00:16:14.360 }, 00:16:14.360 "auth": { 00:16:14.360 "state": "completed", 00:16:14.360 "digest": "sha256", 00:16:14.360 "dhgroup": "ffdhe3072" 00:16:14.360 } 00:16:14.360 } 00:16:14.360 ]' 00:16:14.360 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:14.360 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:14.360 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:14.360 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:16:14.360 18:02:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:14.360 18:02:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:14.360 18:02:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:14.360 18:02:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:14.616 18:02:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:02:M2JjZDM2NWNhZmQ5MTRjN2UwY2Q2YjhiOWUwYmE1MzZiNzNiZDYwNjZhZGJkYjM3IIKM4Q==: --dhchap-ctrl-secret DHHC-1:01:OTFlOTkyN2M2NzZiZWZkNWRjOTE4ZmUxNTUyYjI3ODS55lrp: 00:16:15.178 18:02:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:15.178 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:15.178 18:02:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:15.178 18:02:08 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:15.178 18:02:08 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:15.178 18:02:08 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:15.178 18:02:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:15.178 18:02:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:16:15.178 18:02:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:16:15.435 18:02:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 ffdhe3072 3 00:16:15.435 18:02:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:15.435 18:02:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:15.435 18:02:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe3072 00:16:15.435 18:02:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key3 00:16:15.435 18:02:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:15.435 18:02:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key3 00:16:15.435 18:02:08 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:15.435 18:02:08 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:15.436 18:02:08 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:15.436 18:02:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:16:15.436 18:02:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:16:15.692 00:16:15.692 18:02:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:15.692 18:02:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:15.692 18:02:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:15.692 18:02:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:15.692 18:02:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:15.692 18:02:09 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:15.692 18:02:09 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:15.692 18:02:09 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:15.692 18:02:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:15.692 { 00:16:15.692 "cntlid": 23, 00:16:15.692 "qid": 0, 00:16:15.692 "state": "enabled", 00:16:15.692 "thread": "nvmf_tgt_poll_group_000", 00:16:15.692 "listen_address": { 00:16:15.692 "trtype": "TCP", 00:16:15.692 "adrfam": "IPv4", 00:16:15.692 "traddr": "10.0.0.2", 00:16:15.692 "trsvcid": "4420" 00:16:15.692 }, 00:16:15.692 "peer_address": { 00:16:15.692 "trtype": "TCP", 00:16:15.692 "adrfam": "IPv4", 00:16:15.693 "traddr": "10.0.0.1", 00:16:15.693 "trsvcid": "41110" 00:16:15.693 }, 00:16:15.693 "auth": { 00:16:15.693 "state": "completed", 00:16:15.693 "digest": "sha256", 00:16:15.693 "dhgroup": "ffdhe3072" 00:16:15.693 } 00:16:15.693 } 00:16:15.693 ]' 00:16:15.693 18:02:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:15.950 18:02:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:15.950 18:02:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:15.950 18:02:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:16:15.950 18:02:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:15.950 18:02:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:15.950 18:02:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:15.950 18:02:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:15.950 18:02:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:03:MDkyZTk1N2U3YjVlMDJjNThmOTAwMDMzY2YzZWQ3ZjZiYTk2ZDA3MzczYTM5MTgyNDYxMjY5ZjBjMDlhNGMzZRaivds=: 00:16:16.514 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:16.772 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:16.772 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:16.772 18:02:10 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:16.772 18:02:10 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:16.772 18:02:10 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:16.772 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@92 -- # for dhgroup in "${dhgroups[@]}" 00:16:16.772 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:16.772 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:16:16.772 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:16:16.772 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 ffdhe4096 0 00:16:16.772 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:16.772 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:16.772 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe4096 00:16:16.772 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key0 00:16:16.772 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:16.772 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:16.772 18:02:10 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:16.772 18:02:10 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:16.772 18:02:10 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:16.772 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:16.772 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:17.029 00:16:17.029 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:17.029 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:17.029 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:17.287 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:17.287 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:17.287 18:02:10 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:17.287 18:02:10 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:17.287 18:02:10 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:17.287 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:17.287 { 00:16:17.287 "cntlid": 25, 00:16:17.287 "qid": 0, 00:16:17.287 "state": "enabled", 00:16:17.287 "thread": "nvmf_tgt_poll_group_000", 00:16:17.287 "listen_address": { 00:16:17.287 "trtype": "TCP", 00:16:17.287 "adrfam": "IPv4", 00:16:17.287 "traddr": "10.0.0.2", 00:16:17.287 "trsvcid": "4420" 00:16:17.287 }, 00:16:17.287 "peer_address": { 00:16:17.287 "trtype": "TCP", 00:16:17.287 "adrfam": "IPv4", 00:16:17.287 "traddr": "10.0.0.1", 00:16:17.287 "trsvcid": "41132" 00:16:17.287 }, 00:16:17.287 "auth": { 00:16:17.287 "state": "completed", 00:16:17.287 "digest": "sha256", 00:16:17.287 "dhgroup": "ffdhe4096" 00:16:17.287 } 00:16:17.287 } 00:16:17.287 ]' 00:16:17.287 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:17.287 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:17.287 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:17.287 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:16:17.287 18:02:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:17.543 18:02:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:17.543 18:02:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:17.543 18:02:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:17.543 18:02:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:00:ZDc0NTQyMWE2OWE3NzMyMzhhZTE1NDNiYzY0MmRlNGJhY2Q1NjlmNTk1NGFlN2VlXS98/Q==: --dhchap-ctrl-secret DHHC-1:03:NmM0ODExMWYzNDJlOTA1ZjgyNTliZThhZmQ2NGY1YTMwZGFmYmVkMTY3MmFiNjRkZDdiZTAwNmRhZDYxYmI1N4Bp9x8=: 00:16:18.108 18:02:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:18.108 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:18.108 18:02:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:18.108 18:02:11 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:18.108 18:02:11 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:18.365 18:02:11 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:18.365 18:02:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:18.365 18:02:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:16:18.365 18:02:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:16:18.365 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 ffdhe4096 1 00:16:18.365 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:18.365 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:18.365 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe4096 00:16:18.365 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key1 00:16:18.365 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:18.365 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:18.365 18:02:12 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:18.365 18:02:12 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:18.365 18:02:12 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:18.365 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:18.365 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:18.622 00:16:18.622 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:18.622 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:18.622 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:18.879 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:18.879 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:18.879 18:02:12 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:18.879 18:02:12 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:18.879 18:02:12 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:18.879 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:18.879 { 00:16:18.879 "cntlid": 27, 00:16:18.879 "qid": 0, 00:16:18.879 "state": "enabled", 00:16:18.879 "thread": "nvmf_tgt_poll_group_000", 00:16:18.879 "listen_address": { 00:16:18.879 "trtype": "TCP", 00:16:18.879 "adrfam": "IPv4", 00:16:18.879 "traddr": "10.0.0.2", 00:16:18.879 "trsvcid": "4420" 00:16:18.879 }, 00:16:18.879 "peer_address": { 00:16:18.879 "trtype": "TCP", 00:16:18.879 "adrfam": "IPv4", 00:16:18.879 "traddr": "10.0.0.1", 00:16:18.879 "trsvcid": "41164" 00:16:18.879 }, 00:16:18.879 "auth": { 00:16:18.879 "state": "completed", 00:16:18.879 "digest": "sha256", 00:16:18.879 "dhgroup": "ffdhe4096" 00:16:18.879 } 00:16:18.879 } 00:16:18.879 ]' 00:16:18.879 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:18.879 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:18.879 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:18.879 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:16:18.879 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:18.879 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:18.879 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:18.879 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:19.137 18:02:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:01:ZDMzNjZhODE5ZDkwZjc1YzJlMWRhMzJmNzNkZWZjM2TGg2/J: --dhchap-ctrl-secret DHHC-1:02:OWFmZjc0MDk4ZDJmZWMyYWFjNmJmNmM1MzZkMjUwMDZlY2RmOGMxNjQxNzZkZDdloY1p5Q==: 00:16:19.701 18:02:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:19.701 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:19.701 18:02:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:19.701 18:02:13 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:19.701 18:02:13 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:19.701 18:02:13 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:19.701 18:02:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:19.701 18:02:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:16:19.701 18:02:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:16:19.958 18:02:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 ffdhe4096 2 00:16:19.958 18:02:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:19.958 18:02:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:19.958 18:02:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe4096 00:16:19.958 18:02:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key2 00:16:19.958 18:02:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:19.958 18:02:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:19.958 18:02:13 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:19.958 18:02:13 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:19.958 18:02:13 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:19.958 18:02:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:19.958 18:02:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:20.215 00:16:20.215 18:02:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:20.215 18:02:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:20.215 18:02:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:20.472 18:02:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:20.472 18:02:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:20.472 18:02:13 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:20.472 18:02:13 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:20.472 18:02:14 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:20.472 18:02:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:20.472 { 00:16:20.472 "cntlid": 29, 00:16:20.472 "qid": 0, 00:16:20.472 "state": "enabled", 00:16:20.472 "thread": "nvmf_tgt_poll_group_000", 00:16:20.472 "listen_address": { 00:16:20.472 "trtype": "TCP", 00:16:20.472 "adrfam": "IPv4", 00:16:20.472 "traddr": "10.0.0.2", 00:16:20.472 "trsvcid": "4420" 00:16:20.472 }, 00:16:20.472 "peer_address": { 00:16:20.472 "trtype": "TCP", 00:16:20.472 "adrfam": "IPv4", 00:16:20.472 "traddr": "10.0.0.1", 00:16:20.472 "trsvcid": "41186" 00:16:20.472 }, 00:16:20.472 "auth": { 00:16:20.472 "state": "completed", 00:16:20.472 "digest": "sha256", 00:16:20.472 "dhgroup": "ffdhe4096" 00:16:20.472 } 00:16:20.472 } 00:16:20.472 ]' 00:16:20.472 18:02:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:20.472 18:02:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:20.472 18:02:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:20.472 18:02:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:16:20.472 18:02:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:20.472 18:02:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:20.472 18:02:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:20.472 18:02:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:20.729 18:02:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:02:M2JjZDM2NWNhZmQ5MTRjN2UwY2Q2YjhiOWUwYmE1MzZiNzNiZDYwNjZhZGJkYjM3IIKM4Q==: --dhchap-ctrl-secret DHHC-1:01:OTFlOTkyN2M2NzZiZWZkNWRjOTE4ZmUxNTUyYjI3ODS55lrp: 00:16:21.292 18:02:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:21.292 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:21.292 18:02:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:21.292 18:02:14 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:21.292 18:02:14 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:21.292 18:02:14 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:21.292 18:02:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:21.292 18:02:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:16:21.292 18:02:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:16:21.548 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 ffdhe4096 3 00:16:21.548 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:21.548 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:21.548 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe4096 00:16:21.548 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key3 00:16:21.548 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:21.548 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key3 00:16:21.548 18:02:15 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:21.548 18:02:15 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:21.548 18:02:15 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:21.548 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:16:21.548 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:16:21.804 00:16:21.804 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:21.804 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:21.804 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:21.804 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:21.804 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:21.804 18:02:15 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:21.804 18:02:15 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:21.804 18:02:15 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:22.061 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:22.061 { 00:16:22.061 "cntlid": 31, 00:16:22.061 "qid": 0, 00:16:22.061 "state": "enabled", 00:16:22.061 "thread": "nvmf_tgt_poll_group_000", 00:16:22.061 "listen_address": { 00:16:22.061 "trtype": "TCP", 00:16:22.061 "adrfam": "IPv4", 00:16:22.061 "traddr": "10.0.0.2", 00:16:22.061 "trsvcid": "4420" 00:16:22.061 }, 00:16:22.061 "peer_address": { 00:16:22.061 "trtype": "TCP", 00:16:22.061 "adrfam": "IPv4", 00:16:22.061 "traddr": "10.0.0.1", 00:16:22.061 "trsvcid": "41210" 00:16:22.061 }, 00:16:22.061 "auth": { 00:16:22.061 "state": "completed", 00:16:22.061 "digest": "sha256", 00:16:22.061 "dhgroup": "ffdhe4096" 00:16:22.061 } 00:16:22.061 } 00:16:22.061 ]' 00:16:22.061 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:22.061 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:22.061 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:22.061 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:16:22.061 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:22.061 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:22.061 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:22.061 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:22.317 18:02:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:03:MDkyZTk1N2U3YjVlMDJjNThmOTAwMDMzY2YzZWQ3ZjZiYTk2ZDA3MzczYTM5MTgyNDYxMjY5ZjBjMDlhNGMzZRaivds=: 00:16:22.879 18:02:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:22.879 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:22.879 18:02:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:22.879 18:02:16 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:22.879 18:02:16 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:22.879 18:02:16 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:22.879 18:02:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@92 -- # for dhgroup in "${dhgroups[@]}" 00:16:22.879 18:02:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:22.879 18:02:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:16:22.879 18:02:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:16:22.879 18:02:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 ffdhe6144 0 00:16:22.879 18:02:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:22.879 18:02:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:22.879 18:02:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe6144 00:16:22.879 18:02:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key0 00:16:22.879 18:02:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:22.879 18:02:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:22.879 18:02:16 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:22.879 18:02:16 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:23.136 18:02:16 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:23.136 18:02:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:23.136 18:02:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:23.392 00:16:23.392 18:02:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:23.392 18:02:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:23.392 18:02:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:23.648 18:02:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:23.648 18:02:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:23.648 18:02:17 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:23.648 18:02:17 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:23.648 18:02:17 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:23.648 18:02:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:23.648 { 00:16:23.648 "cntlid": 33, 00:16:23.648 "qid": 0, 00:16:23.648 "state": "enabled", 00:16:23.648 "thread": "nvmf_tgt_poll_group_000", 00:16:23.648 "listen_address": { 00:16:23.648 "trtype": "TCP", 00:16:23.648 "adrfam": "IPv4", 00:16:23.648 "traddr": "10.0.0.2", 00:16:23.648 "trsvcid": "4420" 00:16:23.648 }, 00:16:23.648 "peer_address": { 00:16:23.648 "trtype": "TCP", 00:16:23.648 "adrfam": "IPv4", 00:16:23.648 "traddr": "10.0.0.1", 00:16:23.648 "trsvcid": "45152" 00:16:23.648 }, 00:16:23.648 "auth": { 00:16:23.648 "state": "completed", 00:16:23.648 "digest": "sha256", 00:16:23.648 "dhgroup": "ffdhe6144" 00:16:23.648 } 00:16:23.648 } 00:16:23.648 ]' 00:16:23.648 18:02:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:23.648 18:02:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:23.648 18:02:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:23.648 18:02:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:16:23.648 18:02:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:23.648 18:02:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:23.648 18:02:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:23.648 18:02:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:23.905 18:02:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:00:ZDc0NTQyMWE2OWE3NzMyMzhhZTE1NDNiYzY0MmRlNGJhY2Q1NjlmNTk1NGFlN2VlXS98/Q==: --dhchap-ctrl-secret DHHC-1:03:NmM0ODExMWYzNDJlOTA1ZjgyNTliZThhZmQ2NGY1YTMwZGFmYmVkMTY3MmFiNjRkZDdiZTAwNmRhZDYxYmI1N4Bp9x8=: 00:16:24.468 18:02:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:24.468 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:24.468 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:24.468 18:02:18 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:24.468 18:02:18 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:24.468 18:02:18 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:24.468 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:24.468 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:16:24.468 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:16:24.468 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 ffdhe6144 1 00:16:24.469 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:24.469 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:24.469 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe6144 00:16:24.469 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key1 00:16:24.469 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:24.469 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:24.469 18:02:18 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:24.469 18:02:18 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:24.752 18:02:18 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:24.752 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:24.752 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:25.021 00:16:25.021 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:25.021 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:25.021 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:25.021 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:25.021 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:25.021 18:02:18 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:25.021 18:02:18 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:25.021 18:02:18 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:25.021 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:25.021 { 00:16:25.021 "cntlid": 35, 00:16:25.021 "qid": 0, 00:16:25.021 "state": "enabled", 00:16:25.021 "thread": "nvmf_tgt_poll_group_000", 00:16:25.021 "listen_address": { 00:16:25.021 "trtype": "TCP", 00:16:25.021 "adrfam": "IPv4", 00:16:25.021 "traddr": "10.0.0.2", 00:16:25.021 "trsvcid": "4420" 00:16:25.021 }, 00:16:25.021 "peer_address": { 00:16:25.021 "trtype": "TCP", 00:16:25.021 "adrfam": "IPv4", 00:16:25.021 "traddr": "10.0.0.1", 00:16:25.021 "trsvcid": "45182" 00:16:25.021 }, 00:16:25.021 "auth": { 00:16:25.021 "state": "completed", 00:16:25.021 "digest": "sha256", 00:16:25.021 "dhgroup": "ffdhe6144" 00:16:25.021 } 00:16:25.021 } 00:16:25.021 ]' 00:16:25.021 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:25.278 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:25.278 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:25.278 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:16:25.278 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:25.278 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:25.278 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:25.278 18:02:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:25.535 18:02:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:01:ZDMzNjZhODE5ZDkwZjc1YzJlMWRhMzJmNzNkZWZjM2TGg2/J: --dhchap-ctrl-secret DHHC-1:02:OWFmZjc0MDk4ZDJmZWMyYWFjNmJmNmM1MzZkMjUwMDZlY2RmOGMxNjQxNzZkZDdloY1p5Q==: 00:16:26.100 18:02:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:26.100 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:26.100 18:02:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:26.100 18:02:19 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:26.100 18:02:19 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:26.100 18:02:19 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:26.100 18:02:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:26.100 18:02:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:16:26.100 18:02:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:16:26.100 18:02:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 ffdhe6144 2 00:16:26.100 18:02:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:26.100 18:02:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:26.100 18:02:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe6144 00:16:26.100 18:02:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key2 00:16:26.100 18:02:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:26.100 18:02:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:26.100 18:02:19 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:26.100 18:02:19 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:26.100 18:02:19 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:26.100 18:02:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:26.100 18:02:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:26.665 00:16:26.665 18:02:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:26.665 18:02:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:26.665 18:02:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:26.665 18:02:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:26.665 18:02:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:26.665 18:02:20 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:26.665 18:02:20 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:26.665 18:02:20 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:26.665 18:02:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:26.665 { 00:16:26.665 "cntlid": 37, 00:16:26.665 "qid": 0, 00:16:26.665 "state": "enabled", 00:16:26.665 "thread": "nvmf_tgt_poll_group_000", 00:16:26.665 "listen_address": { 00:16:26.665 "trtype": "TCP", 00:16:26.665 "adrfam": "IPv4", 00:16:26.665 "traddr": "10.0.0.2", 00:16:26.665 "trsvcid": "4420" 00:16:26.665 }, 00:16:26.665 "peer_address": { 00:16:26.665 "trtype": "TCP", 00:16:26.665 "adrfam": "IPv4", 00:16:26.665 "traddr": "10.0.0.1", 00:16:26.665 "trsvcid": "45200" 00:16:26.665 }, 00:16:26.665 "auth": { 00:16:26.665 "state": "completed", 00:16:26.665 "digest": "sha256", 00:16:26.665 "dhgroup": "ffdhe6144" 00:16:26.665 } 00:16:26.665 } 00:16:26.665 ]' 00:16:26.665 18:02:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:26.665 18:02:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:26.666 18:02:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:26.923 18:02:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:16:26.923 18:02:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:26.923 18:02:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:26.923 18:02:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:26.923 18:02:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:26.923 18:02:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:02:M2JjZDM2NWNhZmQ5MTRjN2UwY2Q2YjhiOWUwYmE1MzZiNzNiZDYwNjZhZGJkYjM3IIKM4Q==: --dhchap-ctrl-secret DHHC-1:01:OTFlOTkyN2M2NzZiZWZkNWRjOTE4ZmUxNTUyYjI3ODS55lrp: 00:16:27.487 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:27.487 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:27.487 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:27.487 18:02:21 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:27.487 18:02:21 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:27.487 18:02:21 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:27.487 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:27.487 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:16:27.487 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:16:27.744 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 ffdhe6144 3 00:16:27.744 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:27.744 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:27.744 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe6144 00:16:27.744 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key3 00:16:27.744 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:27.744 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key3 00:16:27.744 18:02:21 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:27.744 18:02:21 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:27.744 18:02:21 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:27.744 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:16:27.744 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:16:28.001 00:16:28.001 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:28.001 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:28.001 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:28.259 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:28.259 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:28.259 18:02:21 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:28.259 18:02:21 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:28.259 18:02:21 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:28.259 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:28.259 { 00:16:28.259 "cntlid": 39, 00:16:28.259 "qid": 0, 00:16:28.259 "state": "enabled", 00:16:28.259 "thread": "nvmf_tgt_poll_group_000", 00:16:28.259 "listen_address": { 00:16:28.259 "trtype": "TCP", 00:16:28.259 "adrfam": "IPv4", 00:16:28.259 "traddr": "10.0.0.2", 00:16:28.259 "trsvcid": "4420" 00:16:28.259 }, 00:16:28.259 "peer_address": { 00:16:28.259 "trtype": "TCP", 00:16:28.259 "adrfam": "IPv4", 00:16:28.259 "traddr": "10.0.0.1", 00:16:28.259 "trsvcid": "45222" 00:16:28.259 }, 00:16:28.259 "auth": { 00:16:28.259 "state": "completed", 00:16:28.259 "digest": "sha256", 00:16:28.259 "dhgroup": "ffdhe6144" 00:16:28.259 } 00:16:28.259 } 00:16:28.259 ]' 00:16:28.259 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:28.259 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:28.259 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:28.259 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:16:28.259 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:28.516 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:28.516 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:28.516 18:02:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:28.517 18:02:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:03:MDkyZTk1N2U3YjVlMDJjNThmOTAwMDMzY2YzZWQ3ZjZiYTk2ZDA3MzczYTM5MTgyNDYxMjY5ZjBjMDlhNGMzZRaivds=: 00:16:29.083 18:02:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:29.083 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:29.083 18:02:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:29.083 18:02:22 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:29.083 18:02:22 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:29.083 18:02:22 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:29.083 18:02:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@92 -- # for dhgroup in "${dhgroups[@]}" 00:16:29.083 18:02:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:29.083 18:02:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:16:29.083 18:02:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:16:29.341 18:02:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 ffdhe8192 0 00:16:29.341 18:02:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:29.341 18:02:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:29.341 18:02:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe8192 00:16:29.341 18:02:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key0 00:16:29.341 18:02:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:29.341 18:02:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:29.341 18:02:22 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:29.341 18:02:22 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:29.341 18:02:22 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:29.341 18:02:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:29.341 18:02:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:29.906 00:16:29.906 18:02:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:29.906 18:02:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:29.906 18:02:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:29.906 18:02:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:29.906 18:02:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:29.906 18:02:23 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:29.906 18:02:23 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:29.906 18:02:23 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:29.906 18:02:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:29.906 { 00:16:29.906 "cntlid": 41, 00:16:29.906 "qid": 0, 00:16:29.906 "state": "enabled", 00:16:29.906 "thread": "nvmf_tgt_poll_group_000", 00:16:29.906 "listen_address": { 00:16:29.906 "trtype": "TCP", 00:16:29.906 "adrfam": "IPv4", 00:16:29.906 "traddr": "10.0.0.2", 00:16:29.906 "trsvcid": "4420" 00:16:29.906 }, 00:16:29.906 "peer_address": { 00:16:29.906 "trtype": "TCP", 00:16:29.906 "adrfam": "IPv4", 00:16:29.906 "traddr": "10.0.0.1", 00:16:29.906 "trsvcid": "45242" 00:16:29.906 }, 00:16:29.906 "auth": { 00:16:29.906 "state": "completed", 00:16:29.906 "digest": "sha256", 00:16:29.906 "dhgroup": "ffdhe8192" 00:16:29.906 } 00:16:29.906 } 00:16:29.906 ]' 00:16:29.906 18:02:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:29.906 18:02:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:29.906 18:02:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:30.163 18:02:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:16:30.163 18:02:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:30.163 18:02:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:30.163 18:02:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:30.163 18:02:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:30.163 18:02:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:00:ZDc0NTQyMWE2OWE3NzMyMzhhZTE1NDNiYzY0MmRlNGJhY2Q1NjlmNTk1NGFlN2VlXS98/Q==: --dhchap-ctrl-secret DHHC-1:03:NmM0ODExMWYzNDJlOTA1ZjgyNTliZThhZmQ2NGY1YTMwZGFmYmVkMTY3MmFiNjRkZDdiZTAwNmRhZDYxYmI1N4Bp9x8=: 00:16:30.728 18:02:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:30.986 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:30.986 18:02:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:30.986 18:02:24 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:30.986 18:02:24 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:30.986 18:02:24 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:30.986 18:02:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:30.986 18:02:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:16:30.986 18:02:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:16:30.986 18:02:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 ffdhe8192 1 00:16:30.986 18:02:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:30.986 18:02:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:30.986 18:02:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe8192 00:16:30.986 18:02:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key1 00:16:30.986 18:02:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:30.986 18:02:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:30.986 18:02:24 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:30.986 18:02:24 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:30.986 18:02:24 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:30.986 18:02:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:30.986 18:02:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:31.550 00:16:31.550 18:02:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:31.550 18:02:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:31.550 18:02:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:31.807 18:02:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:31.807 18:02:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:31.807 18:02:25 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:31.807 18:02:25 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:31.807 18:02:25 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:31.807 18:02:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:31.807 { 00:16:31.807 "cntlid": 43, 00:16:31.807 "qid": 0, 00:16:31.807 "state": "enabled", 00:16:31.807 "thread": "nvmf_tgt_poll_group_000", 00:16:31.807 "listen_address": { 00:16:31.807 "trtype": "TCP", 00:16:31.807 "adrfam": "IPv4", 00:16:31.807 "traddr": "10.0.0.2", 00:16:31.807 "trsvcid": "4420" 00:16:31.807 }, 00:16:31.807 "peer_address": { 00:16:31.807 "trtype": "TCP", 00:16:31.807 "adrfam": "IPv4", 00:16:31.807 "traddr": "10.0.0.1", 00:16:31.807 "trsvcid": "45282" 00:16:31.807 }, 00:16:31.807 "auth": { 00:16:31.807 "state": "completed", 00:16:31.807 "digest": "sha256", 00:16:31.807 "dhgroup": "ffdhe8192" 00:16:31.807 } 00:16:31.807 } 00:16:31.807 ]' 00:16:31.807 18:02:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:31.807 18:02:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:31.807 18:02:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:31.807 18:02:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:16:31.807 18:02:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:31.807 18:02:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:31.807 18:02:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:31.807 18:02:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:32.064 18:02:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:01:ZDMzNjZhODE5ZDkwZjc1YzJlMWRhMzJmNzNkZWZjM2TGg2/J: --dhchap-ctrl-secret DHHC-1:02:OWFmZjc0MDk4ZDJmZWMyYWFjNmJmNmM1MzZkMjUwMDZlY2RmOGMxNjQxNzZkZDdloY1p5Q==: 00:16:32.629 18:02:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:32.629 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:32.629 18:02:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:32.629 18:02:26 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:32.629 18:02:26 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:32.629 18:02:26 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:32.629 18:02:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:32.629 18:02:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:16:32.629 18:02:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:16:32.886 18:02:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 ffdhe8192 2 00:16:32.886 18:02:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:32.886 18:02:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:32.886 18:02:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe8192 00:16:32.886 18:02:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key2 00:16:32.886 18:02:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:32.886 18:02:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:32.886 18:02:26 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:32.886 18:02:26 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:32.886 18:02:26 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:32.886 18:02:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:32.886 18:02:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:33.451 00:16:33.451 18:02:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:33.451 18:02:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:33.451 18:02:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:33.451 18:02:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:33.451 18:02:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:33.451 18:02:27 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:33.451 18:02:27 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:33.451 18:02:27 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:33.451 18:02:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:33.451 { 00:16:33.451 "cntlid": 45, 00:16:33.451 "qid": 0, 00:16:33.451 "state": "enabled", 00:16:33.451 "thread": "nvmf_tgt_poll_group_000", 00:16:33.451 "listen_address": { 00:16:33.451 "trtype": "TCP", 00:16:33.451 "adrfam": "IPv4", 00:16:33.451 "traddr": "10.0.0.2", 00:16:33.451 "trsvcid": "4420" 00:16:33.451 }, 00:16:33.451 "peer_address": { 00:16:33.451 "trtype": "TCP", 00:16:33.451 "adrfam": "IPv4", 00:16:33.451 "traddr": "10.0.0.1", 00:16:33.451 "trsvcid": "45312" 00:16:33.451 }, 00:16:33.451 "auth": { 00:16:33.451 "state": "completed", 00:16:33.451 "digest": "sha256", 00:16:33.451 "dhgroup": "ffdhe8192" 00:16:33.451 } 00:16:33.451 } 00:16:33.451 ]' 00:16:33.451 18:02:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:33.451 18:02:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:33.451 18:02:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:33.709 18:02:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:16:33.709 18:02:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:33.709 18:02:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:33.709 18:02:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:33.709 18:02:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:33.709 18:02:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:02:M2JjZDM2NWNhZmQ5MTRjN2UwY2Q2YjhiOWUwYmE1MzZiNzNiZDYwNjZhZGJkYjM3IIKM4Q==: --dhchap-ctrl-secret DHHC-1:01:OTFlOTkyN2M2NzZiZWZkNWRjOTE4ZmUxNTUyYjI3ODS55lrp: 00:16:34.273 18:02:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:34.273 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:34.273 18:02:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:34.273 18:02:27 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:34.273 18:02:27 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:34.273 18:02:27 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:34.273 18:02:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:34.273 18:02:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:16:34.273 18:02:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:16:34.531 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha256 ffdhe8192 3 00:16:34.531 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:34.531 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha256 00:16:34.531 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe8192 00:16:34.531 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key3 00:16:34.531 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:34.531 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key3 00:16:34.531 18:02:28 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:34.531 18:02:28 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:34.531 18:02:28 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:34.531 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:16:34.531 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:16:35.095 00:16:35.095 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:35.095 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:35.095 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:35.352 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:35.352 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:35.352 18:02:28 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:35.352 18:02:28 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:35.352 18:02:28 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:35.352 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:35.352 { 00:16:35.352 "cntlid": 47, 00:16:35.352 "qid": 0, 00:16:35.352 "state": "enabled", 00:16:35.352 "thread": "nvmf_tgt_poll_group_000", 00:16:35.352 "listen_address": { 00:16:35.352 "trtype": "TCP", 00:16:35.352 "adrfam": "IPv4", 00:16:35.352 "traddr": "10.0.0.2", 00:16:35.352 "trsvcid": "4420" 00:16:35.352 }, 00:16:35.352 "peer_address": { 00:16:35.352 "trtype": "TCP", 00:16:35.352 "adrfam": "IPv4", 00:16:35.352 "traddr": "10.0.0.1", 00:16:35.352 "trsvcid": "39322" 00:16:35.352 }, 00:16:35.352 "auth": { 00:16:35.352 "state": "completed", 00:16:35.352 "digest": "sha256", 00:16:35.352 "dhgroup": "ffdhe8192" 00:16:35.352 } 00:16:35.352 } 00:16:35.352 ]' 00:16:35.352 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:35.352 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha256 == \s\h\a\2\5\6 ]] 00:16:35.352 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:35.352 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:16:35.352 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:35.352 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:35.352 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:35.352 18:02:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:35.610 18:02:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:03:MDkyZTk1N2U3YjVlMDJjNThmOTAwMDMzY2YzZWQ3ZjZiYTk2ZDA3MzczYTM5MTgyNDYxMjY5ZjBjMDlhNGMzZRaivds=: 00:16:36.174 18:02:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:36.174 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:36.174 18:02:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:36.174 18:02:29 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:36.174 18:02:29 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:36.174 18:02:29 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:36.174 18:02:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@91 -- # for digest in "${digests[@]}" 00:16:36.174 18:02:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@92 -- # for dhgroup in "${dhgroups[@]}" 00:16:36.174 18:02:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:36.174 18:02:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:16:36.174 18:02:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:16:36.432 18:02:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 null 0 00:16:36.432 18:02:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:36.432 18:02:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:16:36.432 18:02:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=null 00:16:36.432 18:02:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key0 00:16:36.432 18:02:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:36.432 18:02:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:36.432 18:02:29 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:36.432 18:02:29 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:36.432 18:02:29 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:36.432 18:02:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:36.432 18:02:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:36.432 00:16:36.432 18:02:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:36.432 18:02:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:36.432 18:02:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:36.690 18:02:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:36.690 18:02:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:36.690 18:02:30 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:36.690 18:02:30 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:36.690 18:02:30 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:36.690 18:02:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:36.690 { 00:16:36.690 "cntlid": 49, 00:16:36.690 "qid": 0, 00:16:36.690 "state": "enabled", 00:16:36.690 "thread": "nvmf_tgt_poll_group_000", 00:16:36.690 "listen_address": { 00:16:36.690 "trtype": "TCP", 00:16:36.690 "adrfam": "IPv4", 00:16:36.690 "traddr": "10.0.0.2", 00:16:36.690 "trsvcid": "4420" 00:16:36.690 }, 00:16:36.690 "peer_address": { 00:16:36.690 "trtype": "TCP", 00:16:36.690 "adrfam": "IPv4", 00:16:36.690 "traddr": "10.0.0.1", 00:16:36.690 "trsvcid": "39340" 00:16:36.690 }, 00:16:36.690 "auth": { 00:16:36.690 "state": "completed", 00:16:36.690 "digest": "sha384", 00:16:36.690 "dhgroup": "null" 00:16:36.690 } 00:16:36.690 } 00:16:36.690 ]' 00:16:36.690 18:02:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:36.690 18:02:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:36.690 18:02:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:36.947 18:02:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ null == \n\u\l\l ]] 00:16:36.947 18:02:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:36.947 18:02:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:36.947 18:02:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:36.947 18:02:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:36.947 18:02:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:00:ZDc0NTQyMWE2OWE3NzMyMzhhZTE1NDNiYzY0MmRlNGJhY2Q1NjlmNTk1NGFlN2VlXS98/Q==: --dhchap-ctrl-secret DHHC-1:03:NmM0ODExMWYzNDJlOTA1ZjgyNTliZThhZmQ2NGY1YTMwZGFmYmVkMTY3MmFiNjRkZDdiZTAwNmRhZDYxYmI1N4Bp9x8=: 00:16:37.511 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:37.511 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:37.511 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:37.511 18:02:31 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:37.511 18:02:31 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:37.511 18:02:31 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:37.511 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:37.511 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:16:37.511 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:16:37.769 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 null 1 00:16:37.769 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:37.769 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:16:37.769 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=null 00:16:37.769 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key1 00:16:37.769 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:37.769 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:37.769 18:02:31 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:37.769 18:02:31 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:37.769 18:02:31 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:37.769 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:37.769 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:38.050 00:16:38.050 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:38.050 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:38.050 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:38.307 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:38.307 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:38.307 18:02:31 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:38.307 18:02:31 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:38.307 18:02:31 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:38.307 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:38.307 { 00:16:38.307 "cntlid": 51, 00:16:38.307 "qid": 0, 00:16:38.307 "state": "enabled", 00:16:38.307 "thread": "nvmf_tgt_poll_group_000", 00:16:38.307 "listen_address": { 00:16:38.307 "trtype": "TCP", 00:16:38.307 "adrfam": "IPv4", 00:16:38.307 "traddr": "10.0.0.2", 00:16:38.307 "trsvcid": "4420" 00:16:38.307 }, 00:16:38.307 "peer_address": { 00:16:38.307 "trtype": "TCP", 00:16:38.307 "adrfam": "IPv4", 00:16:38.307 "traddr": "10.0.0.1", 00:16:38.307 "trsvcid": "39360" 00:16:38.307 }, 00:16:38.307 "auth": { 00:16:38.307 "state": "completed", 00:16:38.307 "digest": "sha384", 00:16:38.307 "dhgroup": "null" 00:16:38.307 } 00:16:38.307 } 00:16:38.307 ]' 00:16:38.307 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:38.307 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:38.307 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:38.307 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ null == \n\u\l\l ]] 00:16:38.307 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:38.307 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:38.307 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:38.307 18:02:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:38.564 18:02:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:01:ZDMzNjZhODE5ZDkwZjc1YzJlMWRhMzJmNzNkZWZjM2TGg2/J: --dhchap-ctrl-secret DHHC-1:02:OWFmZjc0MDk4ZDJmZWMyYWFjNmJmNmM1MzZkMjUwMDZlY2RmOGMxNjQxNzZkZDdloY1p5Q==: 00:16:39.149 18:02:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:39.149 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:39.149 18:02:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:39.149 18:02:32 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:39.149 18:02:32 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:39.149 18:02:32 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:39.149 18:02:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:39.149 18:02:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:16:39.149 18:02:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:16:39.405 18:02:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 null 2 00:16:39.405 18:02:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:39.405 18:02:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:16:39.406 18:02:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=null 00:16:39.406 18:02:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key2 00:16:39.406 18:02:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:39.406 18:02:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:39.406 18:02:32 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:39.406 18:02:32 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:39.406 18:02:32 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:39.406 18:02:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:39.406 18:02:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:39.406 00:16:39.406 18:02:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:39.406 18:02:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:39.406 18:02:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:39.661 18:02:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:39.661 18:02:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:39.661 18:02:33 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:39.661 18:02:33 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:39.661 18:02:33 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:39.661 18:02:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:39.661 { 00:16:39.661 "cntlid": 53, 00:16:39.661 "qid": 0, 00:16:39.661 "state": "enabled", 00:16:39.661 "thread": "nvmf_tgt_poll_group_000", 00:16:39.661 "listen_address": { 00:16:39.661 "trtype": "TCP", 00:16:39.661 "adrfam": "IPv4", 00:16:39.661 "traddr": "10.0.0.2", 00:16:39.661 "trsvcid": "4420" 00:16:39.661 }, 00:16:39.661 "peer_address": { 00:16:39.661 "trtype": "TCP", 00:16:39.661 "adrfam": "IPv4", 00:16:39.661 "traddr": "10.0.0.1", 00:16:39.661 "trsvcid": "39390" 00:16:39.661 }, 00:16:39.661 "auth": { 00:16:39.661 "state": "completed", 00:16:39.661 "digest": "sha384", 00:16:39.661 "dhgroup": "null" 00:16:39.661 } 00:16:39.661 } 00:16:39.661 ]' 00:16:39.661 18:02:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:39.661 18:02:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:39.661 18:02:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:39.917 18:02:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ null == \n\u\l\l ]] 00:16:39.917 18:02:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:39.917 18:02:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:39.917 18:02:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:39.917 18:02:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:39.917 18:02:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:02:M2JjZDM2NWNhZmQ5MTRjN2UwY2Q2YjhiOWUwYmE1MzZiNzNiZDYwNjZhZGJkYjM3IIKM4Q==: --dhchap-ctrl-secret DHHC-1:01:OTFlOTkyN2M2NzZiZWZkNWRjOTE4ZmUxNTUyYjI3ODS55lrp: 00:16:40.482 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:40.482 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:40.482 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:40.482 18:02:34 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:40.482 18:02:34 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:40.482 18:02:34 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:40.482 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:40.482 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:16:40.482 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups null 00:16:40.739 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 null 3 00:16:40.739 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:40.739 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:16:40.739 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=null 00:16:40.739 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key3 00:16:40.739 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:40.739 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key3 00:16:40.739 18:02:34 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:40.739 18:02:34 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:40.739 18:02:34 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:40.739 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:16:40.739 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:16:40.996 00:16:40.996 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:40.996 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:40.996 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:41.253 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:41.253 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:41.253 18:02:34 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:41.253 18:02:34 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:41.253 18:02:34 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:41.253 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:41.253 { 00:16:41.253 "cntlid": 55, 00:16:41.253 "qid": 0, 00:16:41.253 "state": "enabled", 00:16:41.253 "thread": "nvmf_tgt_poll_group_000", 00:16:41.253 "listen_address": { 00:16:41.253 "trtype": "TCP", 00:16:41.253 "adrfam": "IPv4", 00:16:41.253 "traddr": "10.0.0.2", 00:16:41.253 "trsvcid": "4420" 00:16:41.253 }, 00:16:41.253 "peer_address": { 00:16:41.253 "trtype": "TCP", 00:16:41.253 "adrfam": "IPv4", 00:16:41.253 "traddr": "10.0.0.1", 00:16:41.253 "trsvcid": "39416" 00:16:41.253 }, 00:16:41.253 "auth": { 00:16:41.253 "state": "completed", 00:16:41.253 "digest": "sha384", 00:16:41.253 "dhgroup": "null" 00:16:41.253 } 00:16:41.253 } 00:16:41.253 ]' 00:16:41.253 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:41.253 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:41.253 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:41.253 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ null == \n\u\l\l ]] 00:16:41.253 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:41.253 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:41.253 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:41.253 18:02:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:41.512 18:02:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:03:MDkyZTk1N2U3YjVlMDJjNThmOTAwMDMzY2YzZWQ3ZjZiYTk2ZDA3MzczYTM5MTgyNDYxMjY5ZjBjMDlhNGMzZRaivds=: 00:16:42.119 18:02:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:42.119 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:42.119 18:02:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:42.119 18:02:35 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:42.119 18:02:35 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:42.119 18:02:35 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:42.119 18:02:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@92 -- # for dhgroup in "${dhgroups[@]}" 00:16:42.119 18:02:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:42.119 18:02:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:16:42.119 18:02:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:16:42.119 18:02:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 ffdhe2048 0 00:16:42.119 18:02:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:42.119 18:02:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:16:42.119 18:02:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe2048 00:16:42.119 18:02:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key0 00:16:42.119 18:02:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:42.119 18:02:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:42.119 18:02:35 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:42.119 18:02:35 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:42.376 18:02:35 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:42.376 18:02:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:42.376 18:02:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:42.376 00:16:42.376 18:02:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:42.376 18:02:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:42.376 18:02:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:42.634 18:02:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:42.634 18:02:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:42.634 18:02:36 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:42.634 18:02:36 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:42.634 18:02:36 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:42.634 18:02:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:42.634 { 00:16:42.634 "cntlid": 57, 00:16:42.634 "qid": 0, 00:16:42.634 "state": "enabled", 00:16:42.634 "thread": "nvmf_tgt_poll_group_000", 00:16:42.634 "listen_address": { 00:16:42.634 "trtype": "TCP", 00:16:42.634 "adrfam": "IPv4", 00:16:42.634 "traddr": "10.0.0.2", 00:16:42.634 "trsvcid": "4420" 00:16:42.634 }, 00:16:42.634 "peer_address": { 00:16:42.634 "trtype": "TCP", 00:16:42.634 "adrfam": "IPv4", 00:16:42.634 "traddr": "10.0.0.1", 00:16:42.634 "trsvcid": "39454" 00:16:42.634 }, 00:16:42.634 "auth": { 00:16:42.634 "state": "completed", 00:16:42.634 "digest": "sha384", 00:16:42.634 "dhgroup": "ffdhe2048" 00:16:42.634 } 00:16:42.634 } 00:16:42.634 ]' 00:16:42.634 18:02:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:42.634 18:02:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:42.634 18:02:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:42.634 18:02:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:16:42.634 18:02:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:42.891 18:02:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:42.891 18:02:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:42.891 18:02:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:42.891 18:02:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:00:ZDc0NTQyMWE2OWE3NzMyMzhhZTE1NDNiYzY0MmRlNGJhY2Q1NjlmNTk1NGFlN2VlXS98/Q==: --dhchap-ctrl-secret DHHC-1:03:NmM0ODExMWYzNDJlOTA1ZjgyNTliZThhZmQ2NGY1YTMwZGFmYmVkMTY3MmFiNjRkZDdiZTAwNmRhZDYxYmI1N4Bp9x8=: 00:16:43.455 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:43.455 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:43.455 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:43.455 18:02:37 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:43.455 18:02:37 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:43.456 18:02:37 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:43.456 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:43.456 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:16:43.456 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:16:43.713 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 ffdhe2048 1 00:16:43.713 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:43.713 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:16:43.713 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe2048 00:16:43.713 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key1 00:16:43.713 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:43.713 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:43.713 18:02:37 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:43.713 18:02:37 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:43.713 18:02:37 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:43.713 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:43.713 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:43.970 00:16:43.970 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:43.970 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:43.970 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:44.227 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:44.227 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:44.227 18:02:37 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:44.227 18:02:37 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:44.227 18:02:37 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:44.227 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:44.227 { 00:16:44.227 "cntlid": 59, 00:16:44.227 "qid": 0, 00:16:44.227 "state": "enabled", 00:16:44.227 "thread": "nvmf_tgt_poll_group_000", 00:16:44.227 "listen_address": { 00:16:44.227 "trtype": "TCP", 00:16:44.227 "adrfam": "IPv4", 00:16:44.227 "traddr": "10.0.0.2", 00:16:44.227 "trsvcid": "4420" 00:16:44.227 }, 00:16:44.227 "peer_address": { 00:16:44.227 "trtype": "TCP", 00:16:44.227 "adrfam": "IPv4", 00:16:44.227 "traddr": "10.0.0.1", 00:16:44.227 "trsvcid": "50512" 00:16:44.227 }, 00:16:44.227 "auth": { 00:16:44.227 "state": "completed", 00:16:44.228 "digest": "sha384", 00:16:44.228 "dhgroup": "ffdhe2048" 00:16:44.228 } 00:16:44.228 } 00:16:44.228 ]' 00:16:44.228 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:44.228 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:44.228 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:44.228 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:16:44.228 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:44.228 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:44.228 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:44.228 18:02:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:44.485 18:02:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:01:ZDMzNjZhODE5ZDkwZjc1YzJlMWRhMzJmNzNkZWZjM2TGg2/J: --dhchap-ctrl-secret DHHC-1:02:OWFmZjc0MDk4ZDJmZWMyYWFjNmJmNmM1MzZkMjUwMDZlY2RmOGMxNjQxNzZkZDdloY1p5Q==: 00:16:45.049 18:02:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:45.049 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:45.049 18:02:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:45.049 18:02:38 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:45.049 18:02:38 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:45.049 18:02:38 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:45.049 18:02:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:45.049 18:02:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:16:45.049 18:02:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:16:45.306 18:02:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 ffdhe2048 2 00:16:45.306 18:02:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:45.306 18:02:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:16:45.306 18:02:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe2048 00:16:45.306 18:02:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key2 00:16:45.307 18:02:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:45.307 18:02:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:45.307 18:02:38 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:45.307 18:02:38 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:45.307 18:02:38 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:45.307 18:02:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:45.307 18:02:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:45.307 00:16:45.307 18:02:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:45.307 18:02:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:45.307 18:02:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:45.564 18:02:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:45.564 18:02:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:45.564 18:02:39 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:45.564 18:02:39 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:45.564 18:02:39 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:45.564 18:02:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:45.564 { 00:16:45.564 "cntlid": 61, 00:16:45.564 "qid": 0, 00:16:45.564 "state": "enabled", 00:16:45.564 "thread": "nvmf_tgt_poll_group_000", 00:16:45.564 "listen_address": { 00:16:45.564 "trtype": "TCP", 00:16:45.564 "adrfam": "IPv4", 00:16:45.564 "traddr": "10.0.0.2", 00:16:45.564 "trsvcid": "4420" 00:16:45.564 }, 00:16:45.564 "peer_address": { 00:16:45.564 "trtype": "TCP", 00:16:45.564 "adrfam": "IPv4", 00:16:45.564 "traddr": "10.0.0.1", 00:16:45.564 "trsvcid": "50526" 00:16:45.564 }, 00:16:45.564 "auth": { 00:16:45.564 "state": "completed", 00:16:45.564 "digest": "sha384", 00:16:45.564 "dhgroup": "ffdhe2048" 00:16:45.564 } 00:16:45.564 } 00:16:45.564 ]' 00:16:45.564 18:02:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:45.564 18:02:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:45.564 18:02:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:45.564 18:02:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:16:45.564 18:02:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:45.822 18:02:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:45.822 18:02:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:45.822 18:02:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:45.822 18:02:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:02:M2JjZDM2NWNhZmQ5MTRjN2UwY2Q2YjhiOWUwYmE1MzZiNzNiZDYwNjZhZGJkYjM3IIKM4Q==: --dhchap-ctrl-secret DHHC-1:01:OTFlOTkyN2M2NzZiZWZkNWRjOTE4ZmUxNTUyYjI3ODS55lrp: 00:16:46.386 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:46.386 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:46.386 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:46.386 18:02:40 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:46.386 18:02:40 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:46.386 18:02:40 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:46.386 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:46.386 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:16:46.386 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:16:46.643 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 ffdhe2048 3 00:16:46.643 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:46.643 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:16:46.643 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe2048 00:16:46.643 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key3 00:16:46.643 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:46.643 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key3 00:16:46.643 18:02:40 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:46.643 18:02:40 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:46.643 18:02:40 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:46.643 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:16:46.643 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:16:46.900 00:16:46.900 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:46.900 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:46.900 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:47.158 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:47.158 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:47.158 18:02:40 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:47.158 18:02:40 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:47.158 18:02:40 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:47.158 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:47.158 { 00:16:47.158 "cntlid": 63, 00:16:47.158 "qid": 0, 00:16:47.158 "state": "enabled", 00:16:47.158 "thread": "nvmf_tgt_poll_group_000", 00:16:47.158 "listen_address": { 00:16:47.158 "trtype": "TCP", 00:16:47.158 "adrfam": "IPv4", 00:16:47.158 "traddr": "10.0.0.2", 00:16:47.158 "trsvcid": "4420" 00:16:47.158 }, 00:16:47.158 "peer_address": { 00:16:47.158 "trtype": "TCP", 00:16:47.158 "adrfam": "IPv4", 00:16:47.158 "traddr": "10.0.0.1", 00:16:47.158 "trsvcid": "50554" 00:16:47.158 }, 00:16:47.158 "auth": { 00:16:47.158 "state": "completed", 00:16:47.158 "digest": "sha384", 00:16:47.158 "dhgroup": "ffdhe2048" 00:16:47.158 } 00:16:47.158 } 00:16:47.158 ]' 00:16:47.158 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:47.158 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:47.158 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:47.158 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:16:47.158 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:47.158 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:47.158 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:47.158 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:47.415 18:02:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:03:MDkyZTk1N2U3YjVlMDJjNThmOTAwMDMzY2YzZWQ3ZjZiYTk2ZDA3MzczYTM5MTgyNDYxMjY5ZjBjMDlhNGMzZRaivds=: 00:16:47.979 18:02:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:47.979 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:47.979 18:02:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:47.979 18:02:41 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:47.979 18:02:41 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:47.979 18:02:41 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:47.979 18:02:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@92 -- # for dhgroup in "${dhgroups[@]}" 00:16:47.979 18:02:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:47.979 18:02:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:16:47.979 18:02:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:16:48.237 18:02:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 ffdhe3072 0 00:16:48.237 18:02:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:48.237 18:02:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:16:48.237 18:02:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe3072 00:16:48.237 18:02:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key0 00:16:48.237 18:02:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:48.237 18:02:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:48.237 18:02:41 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:48.237 18:02:41 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:48.237 18:02:41 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:48.237 18:02:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:48.237 18:02:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:48.237 00:16:48.237 18:02:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:48.237 18:02:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:48.237 18:02:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:48.494 18:02:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:48.494 18:02:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:48.494 18:02:42 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:48.494 18:02:42 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:48.494 18:02:42 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:48.494 18:02:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:48.494 { 00:16:48.494 "cntlid": 65, 00:16:48.494 "qid": 0, 00:16:48.494 "state": "enabled", 00:16:48.494 "thread": "nvmf_tgt_poll_group_000", 00:16:48.494 "listen_address": { 00:16:48.494 "trtype": "TCP", 00:16:48.494 "adrfam": "IPv4", 00:16:48.494 "traddr": "10.0.0.2", 00:16:48.494 "trsvcid": "4420" 00:16:48.494 }, 00:16:48.494 "peer_address": { 00:16:48.494 "trtype": "TCP", 00:16:48.494 "adrfam": "IPv4", 00:16:48.494 "traddr": "10.0.0.1", 00:16:48.495 "trsvcid": "50574" 00:16:48.495 }, 00:16:48.495 "auth": { 00:16:48.495 "state": "completed", 00:16:48.495 "digest": "sha384", 00:16:48.495 "dhgroup": "ffdhe3072" 00:16:48.495 } 00:16:48.495 } 00:16:48.495 ]' 00:16:48.495 18:02:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:48.495 18:02:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:48.495 18:02:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:48.752 18:02:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:16:48.752 18:02:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:48.752 18:02:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:48.752 18:02:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:48.752 18:02:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:48.752 18:02:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:00:ZDc0NTQyMWE2OWE3NzMyMzhhZTE1NDNiYzY0MmRlNGJhY2Q1NjlmNTk1NGFlN2VlXS98/Q==: --dhchap-ctrl-secret DHHC-1:03:NmM0ODExMWYzNDJlOTA1ZjgyNTliZThhZmQ2NGY1YTMwZGFmYmVkMTY3MmFiNjRkZDdiZTAwNmRhZDYxYmI1N4Bp9x8=: 00:16:49.316 18:02:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:49.316 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:49.316 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:49.316 18:02:43 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:49.316 18:02:43 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:49.316 18:02:43 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:49.316 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:49.316 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:16:49.316 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:16:49.573 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 ffdhe3072 1 00:16:49.573 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:49.573 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:16:49.573 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe3072 00:16:49.573 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key1 00:16:49.573 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:49.573 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:49.573 18:02:43 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:49.573 18:02:43 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:49.573 18:02:43 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:49.573 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:49.573 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:49.829 00:16:49.829 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:49.829 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:49.829 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:50.085 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:50.085 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:50.085 18:02:43 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:50.085 18:02:43 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:50.085 18:02:43 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:50.085 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:50.085 { 00:16:50.085 "cntlid": 67, 00:16:50.085 "qid": 0, 00:16:50.085 "state": "enabled", 00:16:50.085 "thread": "nvmf_tgt_poll_group_000", 00:16:50.085 "listen_address": { 00:16:50.085 "trtype": "TCP", 00:16:50.085 "adrfam": "IPv4", 00:16:50.085 "traddr": "10.0.0.2", 00:16:50.085 "trsvcid": "4420" 00:16:50.085 }, 00:16:50.085 "peer_address": { 00:16:50.085 "trtype": "TCP", 00:16:50.085 "adrfam": "IPv4", 00:16:50.085 "traddr": "10.0.0.1", 00:16:50.085 "trsvcid": "50606" 00:16:50.085 }, 00:16:50.085 "auth": { 00:16:50.085 "state": "completed", 00:16:50.085 "digest": "sha384", 00:16:50.085 "dhgroup": "ffdhe3072" 00:16:50.085 } 00:16:50.085 } 00:16:50.085 ]' 00:16:50.085 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:50.085 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:50.085 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:50.085 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:16:50.085 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:50.085 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:50.085 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:50.085 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:50.340 18:02:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:01:ZDMzNjZhODE5ZDkwZjc1YzJlMWRhMzJmNzNkZWZjM2TGg2/J: --dhchap-ctrl-secret DHHC-1:02:OWFmZjc0MDk4ZDJmZWMyYWFjNmJmNmM1MzZkMjUwMDZlY2RmOGMxNjQxNzZkZDdloY1p5Q==: 00:16:50.902 18:02:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:50.902 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:50.902 18:02:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:50.902 18:02:44 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:50.902 18:02:44 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:50.902 18:02:44 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:50.902 18:02:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:50.902 18:02:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:16:50.902 18:02:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:16:51.158 18:02:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 ffdhe3072 2 00:16:51.158 18:02:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:51.158 18:02:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:16:51.159 18:02:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe3072 00:16:51.159 18:02:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key2 00:16:51.159 18:02:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:51.159 18:02:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:51.159 18:02:44 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:51.159 18:02:44 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:51.159 18:02:44 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:51.159 18:02:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:51.159 18:02:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:51.415 00:16:51.415 18:02:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:51.415 18:02:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:51.415 18:02:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:51.415 18:02:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:51.415 18:02:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:51.415 18:02:45 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:51.415 18:02:45 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:51.415 18:02:45 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:51.416 18:02:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:51.416 { 00:16:51.416 "cntlid": 69, 00:16:51.416 "qid": 0, 00:16:51.416 "state": "enabled", 00:16:51.416 "thread": "nvmf_tgt_poll_group_000", 00:16:51.416 "listen_address": { 00:16:51.416 "trtype": "TCP", 00:16:51.416 "adrfam": "IPv4", 00:16:51.416 "traddr": "10.0.0.2", 00:16:51.416 "trsvcid": "4420" 00:16:51.416 }, 00:16:51.416 "peer_address": { 00:16:51.416 "trtype": "TCP", 00:16:51.416 "adrfam": "IPv4", 00:16:51.416 "traddr": "10.0.0.1", 00:16:51.416 "trsvcid": "50638" 00:16:51.416 }, 00:16:51.416 "auth": { 00:16:51.416 "state": "completed", 00:16:51.416 "digest": "sha384", 00:16:51.416 "dhgroup": "ffdhe3072" 00:16:51.416 } 00:16:51.416 } 00:16:51.416 ]' 00:16:51.673 18:02:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:51.673 18:02:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:51.673 18:02:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:51.673 18:02:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:16:51.673 18:02:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:51.673 18:02:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:51.673 18:02:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:51.673 18:02:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:51.930 18:02:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:02:M2JjZDM2NWNhZmQ5MTRjN2UwY2Q2YjhiOWUwYmE1MzZiNzNiZDYwNjZhZGJkYjM3IIKM4Q==: --dhchap-ctrl-secret DHHC-1:01:OTFlOTkyN2M2NzZiZWZkNWRjOTE4ZmUxNTUyYjI3ODS55lrp: 00:16:52.495 18:02:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:52.495 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:52.495 18:02:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:52.495 18:02:45 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:52.495 18:02:45 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:52.495 18:02:45 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:52.495 18:02:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:52.495 18:02:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:16:52.495 18:02:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:16:52.495 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 ffdhe3072 3 00:16:52.495 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:52.495 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:16:52.495 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe3072 00:16:52.495 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key3 00:16:52.495 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:52.495 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key3 00:16:52.495 18:02:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:52.495 18:02:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:52.495 18:02:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:52.495 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:16:52.495 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:16:52.752 00:16:52.752 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:52.753 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:52.753 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:53.010 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:53.010 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:53.010 18:02:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:53.010 18:02:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:53.010 18:02:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:53.010 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:53.010 { 00:16:53.010 "cntlid": 71, 00:16:53.010 "qid": 0, 00:16:53.010 "state": "enabled", 00:16:53.010 "thread": "nvmf_tgt_poll_group_000", 00:16:53.010 "listen_address": { 00:16:53.010 "trtype": "TCP", 00:16:53.010 "adrfam": "IPv4", 00:16:53.010 "traddr": "10.0.0.2", 00:16:53.010 "trsvcid": "4420" 00:16:53.010 }, 00:16:53.010 "peer_address": { 00:16:53.010 "trtype": "TCP", 00:16:53.010 "adrfam": "IPv4", 00:16:53.010 "traddr": "10.0.0.1", 00:16:53.010 "trsvcid": "50668" 00:16:53.010 }, 00:16:53.010 "auth": { 00:16:53.010 "state": "completed", 00:16:53.010 "digest": "sha384", 00:16:53.010 "dhgroup": "ffdhe3072" 00:16:53.010 } 00:16:53.010 } 00:16:53.010 ]' 00:16:53.010 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:53.010 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:53.010 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:53.010 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:16:53.010 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:53.010 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:53.010 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:53.010 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:53.267 18:02:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:03:MDkyZTk1N2U3YjVlMDJjNThmOTAwMDMzY2YzZWQ3ZjZiYTk2ZDA3MzczYTM5MTgyNDYxMjY5ZjBjMDlhNGMzZRaivds=: 00:16:53.831 18:02:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:53.831 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:53.831 18:02:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:53.831 18:02:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:53.831 18:02:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:53.831 18:02:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:53.831 18:02:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@92 -- # for dhgroup in "${dhgroups[@]}" 00:16:53.831 18:02:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:53.831 18:02:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:16:53.831 18:02:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:16:54.089 18:02:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 ffdhe4096 0 00:16:54.089 18:02:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:54.089 18:02:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:16:54.089 18:02:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe4096 00:16:54.089 18:02:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key0 00:16:54.089 18:02:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:54.089 18:02:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:54.089 18:02:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:54.089 18:02:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:54.089 18:02:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:54.089 18:02:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:54.089 18:02:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:16:54.346 00:16:54.346 18:02:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:54.346 18:02:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:54.346 18:02:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:54.634 18:02:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:54.634 18:02:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:54.634 18:02:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:54.634 18:02:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:54.634 18:02:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:54.634 18:02:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:54.634 { 00:16:54.634 "cntlid": 73, 00:16:54.634 "qid": 0, 00:16:54.634 "state": "enabled", 00:16:54.634 "thread": "nvmf_tgt_poll_group_000", 00:16:54.634 "listen_address": { 00:16:54.634 "trtype": "TCP", 00:16:54.634 "adrfam": "IPv4", 00:16:54.634 "traddr": "10.0.0.2", 00:16:54.634 "trsvcid": "4420" 00:16:54.634 }, 00:16:54.634 "peer_address": { 00:16:54.634 "trtype": "TCP", 00:16:54.634 "adrfam": "IPv4", 00:16:54.634 "traddr": "10.0.0.1", 00:16:54.634 "trsvcid": "40498" 00:16:54.634 }, 00:16:54.634 "auth": { 00:16:54.634 "state": "completed", 00:16:54.634 "digest": "sha384", 00:16:54.634 "dhgroup": "ffdhe4096" 00:16:54.634 } 00:16:54.634 } 00:16:54.634 ]' 00:16:54.634 18:02:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:54.634 18:02:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:54.634 18:02:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:54.634 18:02:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:16:54.634 18:02:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:54.634 18:02:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:54.634 18:02:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:54.634 18:02:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:54.892 18:02:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:00:ZDc0NTQyMWE2OWE3NzMyMzhhZTE1NDNiYzY0MmRlNGJhY2Q1NjlmNTk1NGFlN2VlXS98/Q==: --dhchap-ctrl-secret DHHC-1:03:NmM0ODExMWYzNDJlOTA1ZjgyNTliZThhZmQ2NGY1YTMwZGFmYmVkMTY3MmFiNjRkZDdiZTAwNmRhZDYxYmI1N4Bp9x8=: 00:16:55.457 18:02:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:55.457 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:55.457 18:02:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:55.457 18:02:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:55.457 18:02:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:55.457 18:02:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:55.457 18:02:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:55.457 18:02:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:16:55.457 18:02:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:16:55.457 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 ffdhe4096 1 00:16:55.457 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:55.457 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:16:55.457 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe4096 00:16:55.457 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key1 00:16:55.457 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:55.457 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:55.457 18:02:49 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:55.457 18:02:49 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:55.457 18:02:49 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:55.457 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:55.457 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:16:55.715 00:16:55.715 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:55.715 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:55.715 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:55.973 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:55.973 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:55.973 18:02:49 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:55.973 18:02:49 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:55.973 18:02:49 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:55.973 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:55.973 { 00:16:55.973 "cntlid": 75, 00:16:55.973 "qid": 0, 00:16:55.973 "state": "enabled", 00:16:55.973 "thread": "nvmf_tgt_poll_group_000", 00:16:55.973 "listen_address": { 00:16:55.973 "trtype": "TCP", 00:16:55.973 "adrfam": "IPv4", 00:16:55.973 "traddr": "10.0.0.2", 00:16:55.973 "trsvcid": "4420" 00:16:55.973 }, 00:16:55.973 "peer_address": { 00:16:55.973 "trtype": "TCP", 00:16:55.973 "adrfam": "IPv4", 00:16:55.973 "traddr": "10.0.0.1", 00:16:55.973 "trsvcid": "40528" 00:16:55.973 }, 00:16:55.973 "auth": { 00:16:55.973 "state": "completed", 00:16:55.973 "digest": "sha384", 00:16:55.973 "dhgroup": "ffdhe4096" 00:16:55.973 } 00:16:55.973 } 00:16:55.973 ]' 00:16:55.973 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:55.973 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:55.973 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:55.973 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:16:55.973 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:56.231 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:56.231 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:56.231 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:56.231 18:02:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:01:ZDMzNjZhODE5ZDkwZjc1YzJlMWRhMzJmNzNkZWZjM2TGg2/J: --dhchap-ctrl-secret DHHC-1:02:OWFmZjc0MDk4ZDJmZWMyYWFjNmJmNmM1MzZkMjUwMDZlY2RmOGMxNjQxNzZkZDdloY1p5Q==: 00:16:56.795 18:02:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:56.796 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:56.796 18:02:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:56.796 18:02:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:56.796 18:02:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:56.796 18:02:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:56.796 18:02:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:56.796 18:02:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:16:56.796 18:02:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:16:57.053 18:02:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 ffdhe4096 2 00:16:57.053 18:02:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:57.053 18:02:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:16:57.053 18:02:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe4096 00:16:57.053 18:02:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key2 00:16:57.053 18:02:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:57.053 18:02:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:57.053 18:02:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:57.053 18:02:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:57.053 18:02:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:57.053 18:02:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:57.053 18:02:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:16:57.310 00:16:57.310 18:02:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:57.310 18:02:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:57.310 18:02:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:57.567 18:02:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:57.567 18:02:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:57.567 18:02:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:57.567 18:02:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:57.567 18:02:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:57.567 18:02:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:57.567 { 00:16:57.567 "cntlid": 77, 00:16:57.567 "qid": 0, 00:16:57.567 "state": "enabled", 00:16:57.567 "thread": "nvmf_tgt_poll_group_000", 00:16:57.567 "listen_address": { 00:16:57.567 "trtype": "TCP", 00:16:57.567 "adrfam": "IPv4", 00:16:57.567 "traddr": "10.0.0.2", 00:16:57.567 "trsvcid": "4420" 00:16:57.567 }, 00:16:57.567 "peer_address": { 00:16:57.567 "trtype": "TCP", 00:16:57.567 "adrfam": "IPv4", 00:16:57.567 "traddr": "10.0.0.1", 00:16:57.567 "trsvcid": "40564" 00:16:57.567 }, 00:16:57.567 "auth": { 00:16:57.567 "state": "completed", 00:16:57.567 "digest": "sha384", 00:16:57.567 "dhgroup": "ffdhe4096" 00:16:57.567 } 00:16:57.567 } 00:16:57.567 ]' 00:16:57.567 18:02:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:57.567 18:02:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:57.567 18:02:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:57.567 18:02:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:16:57.567 18:02:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:57.567 18:02:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:57.567 18:02:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:57.567 18:02:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:57.824 18:02:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:02:M2JjZDM2NWNhZmQ5MTRjN2UwY2Q2YjhiOWUwYmE1MzZiNzNiZDYwNjZhZGJkYjM3IIKM4Q==: --dhchap-ctrl-secret DHHC-1:01:OTFlOTkyN2M2NzZiZWZkNWRjOTE4ZmUxNTUyYjI3ODS55lrp: 00:16:58.420 18:02:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:58.420 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:58.420 18:02:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:58.420 18:02:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:58.420 18:02:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:58.420 18:02:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:58.420 18:02:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:58.420 18:02:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:16:58.420 18:02:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:16:58.678 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 ffdhe4096 3 00:16:58.678 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:16:58.678 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:16:58.678 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe4096 00:16:58.678 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key3 00:16:58.678 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:16:58.678 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key3 00:16:58.678 18:02:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:58.678 18:02:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:58.678 18:02:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:58.678 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:16:58.678 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:16:58.937 00:16:58.937 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:16:58.937 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:16:58.937 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:16:58.937 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:16:58.937 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:16:58.937 18:02:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:58.937 18:02:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:58.937 18:02:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:58.937 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:16:58.937 { 00:16:58.937 "cntlid": 79, 00:16:58.937 "qid": 0, 00:16:58.937 "state": "enabled", 00:16:58.937 "thread": "nvmf_tgt_poll_group_000", 00:16:58.937 "listen_address": { 00:16:58.937 "trtype": "TCP", 00:16:58.937 "adrfam": "IPv4", 00:16:58.937 "traddr": "10.0.0.2", 00:16:58.937 "trsvcid": "4420" 00:16:58.937 }, 00:16:58.937 "peer_address": { 00:16:58.937 "trtype": "TCP", 00:16:58.937 "adrfam": "IPv4", 00:16:58.937 "traddr": "10.0.0.1", 00:16:58.937 "trsvcid": "40588" 00:16:58.937 }, 00:16:58.937 "auth": { 00:16:58.937 "state": "completed", 00:16:58.937 "digest": "sha384", 00:16:58.937 "dhgroup": "ffdhe4096" 00:16:58.937 } 00:16:58.937 } 00:16:58.937 ]' 00:16:58.937 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:16:58.937 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:16:58.937 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:16:59.195 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:16:59.195 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:16:59.195 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:16:59.195 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:16:59.195 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:16:59.195 18:02:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:03:MDkyZTk1N2U3YjVlMDJjNThmOTAwMDMzY2YzZWQ3ZjZiYTk2ZDA3MzczYTM5MTgyNDYxMjY5ZjBjMDlhNGMzZRaivds=: 00:16:59.762 18:02:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:16:59.762 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:16:59.762 18:02:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:16:59.762 18:02:53 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:16:59.762 18:02:53 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:16:59.762 18:02:53 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:16:59.762 18:02:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@92 -- # for dhgroup in "${dhgroups[@]}" 00:16:59.762 18:02:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:16:59.762 18:02:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:16:59.762 18:02:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:17:00.020 18:02:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 ffdhe6144 0 00:17:00.020 18:02:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:00.020 18:02:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:17:00.020 18:02:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe6144 00:17:00.020 18:02:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key0 00:17:00.020 18:02:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:00.020 18:02:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:00.020 18:02:53 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:00.020 18:02:53 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:00.020 18:02:53 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:00.020 18:02:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:00.020 18:02:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:00.278 00:17:00.278 18:02:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:00.278 18:02:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:00.278 18:02:53 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:00.536 18:02:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:00.536 18:02:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:00.536 18:02:54 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:00.536 18:02:54 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:00.536 18:02:54 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:00.536 18:02:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:00.536 { 00:17:00.536 "cntlid": 81, 00:17:00.536 "qid": 0, 00:17:00.536 "state": "enabled", 00:17:00.536 "thread": "nvmf_tgt_poll_group_000", 00:17:00.536 "listen_address": { 00:17:00.536 "trtype": "TCP", 00:17:00.536 "adrfam": "IPv4", 00:17:00.536 "traddr": "10.0.0.2", 00:17:00.536 "trsvcid": "4420" 00:17:00.536 }, 00:17:00.536 "peer_address": { 00:17:00.536 "trtype": "TCP", 00:17:00.536 "adrfam": "IPv4", 00:17:00.536 "traddr": "10.0.0.1", 00:17:00.536 "trsvcid": "40620" 00:17:00.536 }, 00:17:00.536 "auth": { 00:17:00.536 "state": "completed", 00:17:00.536 "digest": "sha384", 00:17:00.536 "dhgroup": "ffdhe6144" 00:17:00.536 } 00:17:00.536 } 00:17:00.536 ]' 00:17:00.536 18:02:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:00.536 18:02:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:00.536 18:02:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:00.795 18:02:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:17:00.795 18:02:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:00.795 18:02:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:00.795 18:02:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:00.795 18:02:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:00.795 18:02:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:00:ZDc0NTQyMWE2OWE3NzMyMzhhZTE1NDNiYzY0MmRlNGJhY2Q1NjlmNTk1NGFlN2VlXS98/Q==: --dhchap-ctrl-secret DHHC-1:03:NmM0ODExMWYzNDJlOTA1ZjgyNTliZThhZmQ2NGY1YTMwZGFmYmVkMTY3MmFiNjRkZDdiZTAwNmRhZDYxYmI1N4Bp9x8=: 00:17:01.362 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:01.362 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:01.362 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:01.362 18:02:55 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:01.362 18:02:55 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:01.362 18:02:55 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:01.362 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:01.362 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:17:01.362 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:17:01.620 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 ffdhe6144 1 00:17:01.620 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:01.620 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:17:01.620 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe6144 00:17:01.620 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key1 00:17:01.620 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:01.620 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:01.620 18:02:55 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:01.620 18:02:55 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:01.620 18:02:55 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:01.620 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:01.620 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:01.878 00:17:01.878 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:01.878 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:01.878 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:02.137 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:02.137 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:02.137 18:02:55 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:02.137 18:02:55 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:02.137 18:02:55 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:02.137 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:02.137 { 00:17:02.137 "cntlid": 83, 00:17:02.137 "qid": 0, 00:17:02.137 "state": "enabled", 00:17:02.137 "thread": "nvmf_tgt_poll_group_000", 00:17:02.137 "listen_address": { 00:17:02.137 "trtype": "TCP", 00:17:02.137 "adrfam": "IPv4", 00:17:02.137 "traddr": "10.0.0.2", 00:17:02.137 "trsvcid": "4420" 00:17:02.137 }, 00:17:02.137 "peer_address": { 00:17:02.137 "trtype": "TCP", 00:17:02.137 "adrfam": "IPv4", 00:17:02.137 "traddr": "10.0.0.1", 00:17:02.137 "trsvcid": "40650" 00:17:02.137 }, 00:17:02.137 "auth": { 00:17:02.137 "state": "completed", 00:17:02.137 "digest": "sha384", 00:17:02.137 "dhgroup": "ffdhe6144" 00:17:02.137 } 00:17:02.137 } 00:17:02.137 ]' 00:17:02.137 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:02.137 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:02.137 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:02.395 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:17:02.395 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:02.395 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:02.395 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:02.395 18:02:55 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:02.395 18:02:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:01:ZDMzNjZhODE5ZDkwZjc1YzJlMWRhMzJmNzNkZWZjM2TGg2/J: --dhchap-ctrl-secret DHHC-1:02:OWFmZjc0MDk4ZDJmZWMyYWFjNmJmNmM1MzZkMjUwMDZlY2RmOGMxNjQxNzZkZDdloY1p5Q==: 00:17:02.962 18:02:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:02.962 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:02.962 18:02:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:02.962 18:02:56 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:02.962 18:02:56 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:02.962 18:02:56 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:02.962 18:02:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:02.962 18:02:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:17:02.962 18:02:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:17:03.221 18:02:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 ffdhe6144 2 00:17:03.221 18:02:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:03.221 18:02:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:17:03.221 18:02:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe6144 00:17:03.221 18:02:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key2 00:17:03.221 18:02:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:03.221 18:02:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:03.221 18:02:56 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:03.221 18:02:56 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:03.221 18:02:56 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:03.221 18:02:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:03.221 18:02:56 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:03.480 00:17:03.480 18:02:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:03.480 18:02:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:03.480 18:02:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:03.739 18:02:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:03.739 18:02:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:03.739 18:02:57 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:03.739 18:02:57 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:03.739 18:02:57 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:03.739 18:02:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:03.739 { 00:17:03.739 "cntlid": 85, 00:17:03.739 "qid": 0, 00:17:03.739 "state": "enabled", 00:17:03.739 "thread": "nvmf_tgt_poll_group_000", 00:17:03.739 "listen_address": { 00:17:03.739 "trtype": "TCP", 00:17:03.739 "adrfam": "IPv4", 00:17:03.739 "traddr": "10.0.0.2", 00:17:03.739 "trsvcid": "4420" 00:17:03.739 }, 00:17:03.739 "peer_address": { 00:17:03.739 "trtype": "TCP", 00:17:03.739 "adrfam": "IPv4", 00:17:03.739 "traddr": "10.0.0.1", 00:17:03.739 "trsvcid": "33898" 00:17:03.739 }, 00:17:03.739 "auth": { 00:17:03.739 "state": "completed", 00:17:03.739 "digest": "sha384", 00:17:03.739 "dhgroup": "ffdhe6144" 00:17:03.739 } 00:17:03.739 } 00:17:03.739 ]' 00:17:03.739 18:02:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:03.739 18:02:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:03.739 18:02:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:03.739 18:02:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:17:03.740 18:02:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:03.998 18:02:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:03.998 18:02:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:03.998 18:02:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:03.998 18:02:57 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:02:M2JjZDM2NWNhZmQ5MTRjN2UwY2Q2YjhiOWUwYmE1MzZiNzNiZDYwNjZhZGJkYjM3IIKM4Q==: --dhchap-ctrl-secret DHHC-1:01:OTFlOTkyN2M2NzZiZWZkNWRjOTE4ZmUxNTUyYjI3ODS55lrp: 00:17:04.565 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:04.565 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:04.565 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:04.565 18:02:58 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:04.566 18:02:58 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:04.566 18:02:58 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:04.566 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:04.566 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:17:04.566 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:17:04.824 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 ffdhe6144 3 00:17:04.824 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:04.824 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:17:04.824 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe6144 00:17:04.824 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key3 00:17:04.824 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:04.824 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key3 00:17:04.824 18:02:58 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:04.824 18:02:58 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:04.824 18:02:58 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:04.824 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:04.824 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:05.083 00:17:05.083 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:05.083 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:05.083 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:05.341 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:05.341 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:05.341 18:02:58 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:05.341 18:02:58 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:05.341 18:02:58 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:05.341 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:05.341 { 00:17:05.341 "cntlid": 87, 00:17:05.341 "qid": 0, 00:17:05.341 "state": "enabled", 00:17:05.341 "thread": "nvmf_tgt_poll_group_000", 00:17:05.341 "listen_address": { 00:17:05.341 "trtype": "TCP", 00:17:05.341 "adrfam": "IPv4", 00:17:05.341 "traddr": "10.0.0.2", 00:17:05.341 "trsvcid": "4420" 00:17:05.341 }, 00:17:05.341 "peer_address": { 00:17:05.341 "trtype": "TCP", 00:17:05.341 "adrfam": "IPv4", 00:17:05.341 "traddr": "10.0.0.1", 00:17:05.341 "trsvcid": "33920" 00:17:05.341 }, 00:17:05.341 "auth": { 00:17:05.341 "state": "completed", 00:17:05.341 "digest": "sha384", 00:17:05.341 "dhgroup": "ffdhe6144" 00:17:05.341 } 00:17:05.341 } 00:17:05.341 ]' 00:17:05.341 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:05.341 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:05.341 18:02:58 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:05.341 18:02:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:17:05.341 18:02:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:05.600 18:02:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:05.600 18:02:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:05.600 18:02:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:05.600 18:02:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:03:MDkyZTk1N2U3YjVlMDJjNThmOTAwMDMzY2YzZWQ3ZjZiYTk2ZDA3MzczYTM5MTgyNDYxMjY5ZjBjMDlhNGMzZRaivds=: 00:17:06.166 18:02:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:06.166 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:06.166 18:02:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:06.166 18:02:59 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:06.166 18:02:59 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:06.166 18:02:59 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:06.167 18:02:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@92 -- # for dhgroup in "${dhgroups[@]}" 00:17:06.167 18:02:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:06.167 18:02:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:17:06.167 18:02:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:17:06.426 18:02:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 ffdhe8192 0 00:17:06.426 18:02:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:06.426 18:02:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:17:06.426 18:02:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe8192 00:17:06.426 18:02:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key0 00:17:06.426 18:02:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:06.426 18:02:59 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:06.426 18:02:59 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:06.426 18:02:59 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:06.426 18:03:00 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:06.426 18:03:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:06.426 18:03:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:06.993 00:17:06.993 18:03:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:06.993 18:03:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:06.993 18:03:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:06.993 18:03:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:06.993 18:03:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:06.993 18:03:00 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:06.993 18:03:00 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:06.993 18:03:00 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:06.993 18:03:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:06.993 { 00:17:06.993 "cntlid": 89, 00:17:06.993 "qid": 0, 00:17:06.993 "state": "enabled", 00:17:06.993 "thread": "nvmf_tgt_poll_group_000", 00:17:06.993 "listen_address": { 00:17:06.993 "trtype": "TCP", 00:17:06.993 "adrfam": "IPv4", 00:17:06.993 "traddr": "10.0.0.2", 00:17:06.993 "trsvcid": "4420" 00:17:06.993 }, 00:17:06.993 "peer_address": { 00:17:06.993 "trtype": "TCP", 00:17:06.993 "adrfam": "IPv4", 00:17:06.993 "traddr": "10.0.0.1", 00:17:06.993 "trsvcid": "33948" 00:17:06.993 }, 00:17:06.993 "auth": { 00:17:06.993 "state": "completed", 00:17:06.993 "digest": "sha384", 00:17:06.993 "dhgroup": "ffdhe8192" 00:17:06.993 } 00:17:06.993 } 00:17:06.993 ]' 00:17:06.993 18:03:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:07.251 18:03:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:07.251 18:03:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:07.251 18:03:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:07.251 18:03:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:07.251 18:03:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:07.251 18:03:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:07.251 18:03:00 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:07.509 18:03:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:00:ZDc0NTQyMWE2OWE3NzMyMzhhZTE1NDNiYzY0MmRlNGJhY2Q1NjlmNTk1NGFlN2VlXS98/Q==: --dhchap-ctrl-secret DHHC-1:03:NmM0ODExMWYzNDJlOTA1ZjgyNTliZThhZmQ2NGY1YTMwZGFmYmVkMTY3MmFiNjRkZDdiZTAwNmRhZDYxYmI1N4Bp9x8=: 00:17:08.075 18:03:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:08.075 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:08.075 18:03:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:08.075 18:03:01 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:08.075 18:03:01 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:08.075 18:03:01 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:08.075 18:03:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:08.075 18:03:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:17:08.075 18:03:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:17:08.075 18:03:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 ffdhe8192 1 00:17:08.075 18:03:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:08.075 18:03:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:17:08.075 18:03:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe8192 00:17:08.075 18:03:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key1 00:17:08.075 18:03:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:08.075 18:03:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:08.075 18:03:01 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:08.075 18:03:01 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:08.075 18:03:01 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:08.075 18:03:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:08.075 18:03:01 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:08.642 00:17:08.642 18:03:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:08.642 18:03:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:08.642 18:03:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:08.902 18:03:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:08.902 18:03:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:08.902 18:03:02 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:08.902 18:03:02 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:08.902 18:03:02 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:08.902 18:03:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:08.902 { 00:17:08.902 "cntlid": 91, 00:17:08.902 "qid": 0, 00:17:08.902 "state": "enabled", 00:17:08.902 "thread": "nvmf_tgt_poll_group_000", 00:17:08.902 "listen_address": { 00:17:08.902 "trtype": "TCP", 00:17:08.902 "adrfam": "IPv4", 00:17:08.902 "traddr": "10.0.0.2", 00:17:08.902 "trsvcid": "4420" 00:17:08.902 }, 00:17:08.902 "peer_address": { 00:17:08.902 "trtype": "TCP", 00:17:08.902 "adrfam": "IPv4", 00:17:08.902 "traddr": "10.0.0.1", 00:17:08.902 "trsvcid": "33984" 00:17:08.902 }, 00:17:08.902 "auth": { 00:17:08.902 "state": "completed", 00:17:08.902 "digest": "sha384", 00:17:08.902 "dhgroup": "ffdhe8192" 00:17:08.902 } 00:17:08.902 } 00:17:08.902 ]' 00:17:08.902 18:03:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:08.902 18:03:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:08.902 18:03:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:08.902 18:03:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:08.902 18:03:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:08.902 18:03:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:08.902 18:03:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:08.902 18:03:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:09.161 18:03:02 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:01:ZDMzNjZhODE5ZDkwZjc1YzJlMWRhMzJmNzNkZWZjM2TGg2/J: --dhchap-ctrl-secret DHHC-1:02:OWFmZjc0MDk4ZDJmZWMyYWFjNmJmNmM1MzZkMjUwMDZlY2RmOGMxNjQxNzZkZDdloY1p5Q==: 00:17:09.728 18:03:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:09.728 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:09.728 18:03:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:09.728 18:03:03 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:09.728 18:03:03 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:09.728 18:03:03 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:09.728 18:03:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:09.728 18:03:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:17:09.728 18:03:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:17:09.987 18:03:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 ffdhe8192 2 00:17:09.987 18:03:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:09.987 18:03:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:17:09.987 18:03:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe8192 00:17:09.987 18:03:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key2 00:17:09.987 18:03:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:09.987 18:03:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:09.987 18:03:03 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:09.987 18:03:03 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:09.987 18:03:03 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:09.987 18:03:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:09.987 18:03:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:10.555 00:17:10.555 18:03:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:10.555 18:03:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:10.555 18:03:03 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:10.555 18:03:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:10.555 18:03:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:10.555 18:03:04 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:10.555 18:03:04 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:10.555 18:03:04 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:10.555 18:03:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:10.555 { 00:17:10.555 "cntlid": 93, 00:17:10.555 "qid": 0, 00:17:10.555 "state": "enabled", 00:17:10.555 "thread": "nvmf_tgt_poll_group_000", 00:17:10.555 "listen_address": { 00:17:10.555 "trtype": "TCP", 00:17:10.555 "adrfam": "IPv4", 00:17:10.555 "traddr": "10.0.0.2", 00:17:10.555 "trsvcid": "4420" 00:17:10.555 }, 00:17:10.555 "peer_address": { 00:17:10.555 "trtype": "TCP", 00:17:10.555 "adrfam": "IPv4", 00:17:10.555 "traddr": "10.0.0.1", 00:17:10.555 "trsvcid": "34020" 00:17:10.555 }, 00:17:10.555 "auth": { 00:17:10.555 "state": "completed", 00:17:10.555 "digest": "sha384", 00:17:10.555 "dhgroup": "ffdhe8192" 00:17:10.555 } 00:17:10.555 } 00:17:10.555 ]' 00:17:10.555 18:03:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:10.555 18:03:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:10.555 18:03:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:10.555 18:03:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:10.555 18:03:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:10.814 18:03:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:10.814 18:03:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:10.814 18:03:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:10.814 18:03:04 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:02:M2JjZDM2NWNhZmQ5MTRjN2UwY2Q2YjhiOWUwYmE1MzZiNzNiZDYwNjZhZGJkYjM3IIKM4Q==: --dhchap-ctrl-secret DHHC-1:01:OTFlOTkyN2M2NzZiZWZkNWRjOTE4ZmUxNTUyYjI3ODS55lrp: 00:17:11.381 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:11.381 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:11.381 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:11.381 18:03:05 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:11.381 18:03:05 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:11.381 18:03:05 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:11.381 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:11.381 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:17:11.381 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:17:11.639 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha384 ffdhe8192 3 00:17:11.639 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:11.639 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha384 00:17:11.639 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe8192 00:17:11.639 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key3 00:17:11.639 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:11.639 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key3 00:17:11.639 18:03:05 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:11.639 18:03:05 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:11.639 18:03:05 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:11.639 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:11.639 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:12.207 00:17:12.207 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:12.207 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:12.207 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:12.207 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:12.207 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:12.207 18:03:05 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:12.207 18:03:05 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:12.207 18:03:05 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:12.207 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:12.207 { 00:17:12.207 "cntlid": 95, 00:17:12.207 "qid": 0, 00:17:12.207 "state": "enabled", 00:17:12.207 "thread": "nvmf_tgt_poll_group_000", 00:17:12.207 "listen_address": { 00:17:12.207 "trtype": "TCP", 00:17:12.207 "adrfam": "IPv4", 00:17:12.207 "traddr": "10.0.0.2", 00:17:12.207 "trsvcid": "4420" 00:17:12.207 }, 00:17:12.207 "peer_address": { 00:17:12.207 "trtype": "TCP", 00:17:12.207 "adrfam": "IPv4", 00:17:12.207 "traddr": "10.0.0.1", 00:17:12.207 "trsvcid": "34066" 00:17:12.207 }, 00:17:12.207 "auth": { 00:17:12.207 "state": "completed", 00:17:12.207 "digest": "sha384", 00:17:12.207 "dhgroup": "ffdhe8192" 00:17:12.207 } 00:17:12.207 } 00:17:12.207 ]' 00:17:12.207 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:12.465 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha384 == \s\h\a\3\8\4 ]] 00:17:12.465 18:03:05 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:12.465 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:12.465 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:12.465 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:12.465 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:12.465 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:12.724 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:03:MDkyZTk1N2U3YjVlMDJjNThmOTAwMDMzY2YzZWQ3ZjZiYTk2ZDA3MzczYTM5MTgyNDYxMjY5ZjBjMDlhNGMzZRaivds=: 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:13.296 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@91 -- # for digest in "${digests[@]}" 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@92 -- # for dhgroup in "${dhgroups[@]}" 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 null 0 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=null 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key0 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:13.296 18:03:06 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:13.555 00:17:13.555 18:03:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:13.555 18:03:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:13.555 18:03:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:13.814 18:03:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:13.814 18:03:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:13.814 18:03:07 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:13.814 18:03:07 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:13.814 18:03:07 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:13.814 18:03:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:13.814 { 00:17:13.814 "cntlid": 97, 00:17:13.814 "qid": 0, 00:17:13.814 "state": "enabled", 00:17:13.814 "thread": "nvmf_tgt_poll_group_000", 00:17:13.814 "listen_address": { 00:17:13.814 "trtype": "TCP", 00:17:13.814 "adrfam": "IPv4", 00:17:13.814 "traddr": "10.0.0.2", 00:17:13.814 "trsvcid": "4420" 00:17:13.814 }, 00:17:13.814 "peer_address": { 00:17:13.814 "trtype": "TCP", 00:17:13.814 "adrfam": "IPv4", 00:17:13.814 "traddr": "10.0.0.1", 00:17:13.814 "trsvcid": "59544" 00:17:13.814 }, 00:17:13.814 "auth": { 00:17:13.814 "state": "completed", 00:17:13.814 "digest": "sha512", 00:17:13.814 "dhgroup": "null" 00:17:13.814 } 00:17:13.814 } 00:17:13.814 ]' 00:17:13.814 18:03:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:13.814 18:03:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:13.814 18:03:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:13.814 18:03:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ null == \n\u\l\l ]] 00:17:13.814 18:03:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:13.814 18:03:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:13.814 18:03:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:13.814 18:03:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:14.072 18:03:07 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:00:ZDc0NTQyMWE2OWE3NzMyMzhhZTE1NDNiYzY0MmRlNGJhY2Q1NjlmNTk1NGFlN2VlXS98/Q==: --dhchap-ctrl-secret DHHC-1:03:NmM0ODExMWYzNDJlOTA1ZjgyNTliZThhZmQ2NGY1YTMwZGFmYmVkMTY3MmFiNjRkZDdiZTAwNmRhZDYxYmI1N4Bp9x8=: 00:17:14.640 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:14.640 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:14.640 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:14.640 18:03:08 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:14.640 18:03:08 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:14.640 18:03:08 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:14.640 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:14.640 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:17:14.640 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:17:14.899 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 null 1 00:17:14.899 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:14.899 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:14.899 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=null 00:17:14.899 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key1 00:17:14.899 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:14.899 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:14.899 18:03:08 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:14.899 18:03:08 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:14.899 18:03:08 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:14.899 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:14.899 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:15.189 00:17:15.189 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:15.189 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:15.189 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:15.189 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:15.189 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:15.189 18:03:08 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:15.189 18:03:08 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:15.447 18:03:08 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:15.447 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:15.447 { 00:17:15.447 "cntlid": 99, 00:17:15.447 "qid": 0, 00:17:15.447 "state": "enabled", 00:17:15.447 "thread": "nvmf_tgt_poll_group_000", 00:17:15.447 "listen_address": { 00:17:15.447 "trtype": "TCP", 00:17:15.447 "adrfam": "IPv4", 00:17:15.447 "traddr": "10.0.0.2", 00:17:15.447 "trsvcid": "4420" 00:17:15.447 }, 00:17:15.447 "peer_address": { 00:17:15.447 "trtype": "TCP", 00:17:15.447 "adrfam": "IPv4", 00:17:15.447 "traddr": "10.0.0.1", 00:17:15.447 "trsvcid": "59572" 00:17:15.447 }, 00:17:15.447 "auth": { 00:17:15.447 "state": "completed", 00:17:15.447 "digest": "sha512", 00:17:15.447 "dhgroup": "null" 00:17:15.447 } 00:17:15.447 } 00:17:15.447 ]' 00:17:15.447 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:15.447 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:15.447 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:15.447 18:03:08 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ null == \n\u\l\l ]] 00:17:15.447 18:03:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:15.447 18:03:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:15.447 18:03:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:15.447 18:03:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:15.705 18:03:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:01:ZDMzNjZhODE5ZDkwZjc1YzJlMWRhMzJmNzNkZWZjM2TGg2/J: --dhchap-ctrl-secret DHHC-1:02:OWFmZjc0MDk4ZDJmZWMyYWFjNmJmNmM1MzZkMjUwMDZlY2RmOGMxNjQxNzZkZDdloY1p5Q==: 00:17:16.272 18:03:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:16.272 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:16.272 18:03:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:16.272 18:03:09 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:16.272 18:03:09 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:16.272 18:03:09 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:16.272 18:03:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:16.272 18:03:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:17:16.272 18:03:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:17:16.272 18:03:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 null 2 00:17:16.272 18:03:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:16.272 18:03:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:16.272 18:03:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=null 00:17:16.272 18:03:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key2 00:17:16.272 18:03:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:16.272 18:03:09 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:16.272 18:03:09 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:16.272 18:03:09 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:16.530 18:03:10 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:16.530 18:03:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:16.530 18:03:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:16.530 00:17:16.530 18:03:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:16.530 18:03:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:16.530 18:03:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:16.789 18:03:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:16.789 18:03:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:16.789 18:03:10 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:16.789 18:03:10 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:16.789 18:03:10 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:16.789 18:03:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:16.789 { 00:17:16.789 "cntlid": 101, 00:17:16.789 "qid": 0, 00:17:16.789 "state": "enabled", 00:17:16.789 "thread": "nvmf_tgt_poll_group_000", 00:17:16.789 "listen_address": { 00:17:16.789 "trtype": "TCP", 00:17:16.789 "adrfam": "IPv4", 00:17:16.789 "traddr": "10.0.0.2", 00:17:16.789 "trsvcid": "4420" 00:17:16.789 }, 00:17:16.789 "peer_address": { 00:17:16.789 "trtype": "TCP", 00:17:16.789 "adrfam": "IPv4", 00:17:16.789 "traddr": "10.0.0.1", 00:17:16.789 "trsvcid": "59608" 00:17:16.789 }, 00:17:16.789 "auth": { 00:17:16.789 "state": "completed", 00:17:16.789 "digest": "sha512", 00:17:16.789 "dhgroup": "null" 00:17:16.789 } 00:17:16.789 } 00:17:16.789 ]' 00:17:16.789 18:03:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:16.789 18:03:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:16.789 18:03:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:16.789 18:03:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ null == \n\u\l\l ]] 00:17:16.789 18:03:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:17.047 18:03:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:17.047 18:03:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:17.048 18:03:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:17.048 18:03:10 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:02:M2JjZDM2NWNhZmQ5MTRjN2UwY2Q2YjhiOWUwYmE1MzZiNzNiZDYwNjZhZGJkYjM3IIKM4Q==: --dhchap-ctrl-secret DHHC-1:01:OTFlOTkyN2M2NzZiZWZkNWRjOTE4ZmUxNTUyYjI3ODS55lrp: 00:17:17.614 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:17.614 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:17.615 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:17.615 18:03:11 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:17.615 18:03:11 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:17.615 18:03:11 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:17.615 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:17.615 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:17:17.615 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups null 00:17:17.873 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 null 3 00:17:17.873 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:17.873 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:17.873 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=null 00:17:17.873 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key3 00:17:17.873 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:17.873 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key3 00:17:17.873 18:03:11 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:17.873 18:03:11 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:17.873 18:03:11 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:17.873 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:17.873 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:18.150 00:17:18.150 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:18.150 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:18.150 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:18.410 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:18.410 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:18.410 18:03:11 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:18.410 18:03:11 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:18.410 18:03:11 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:18.410 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:18.410 { 00:17:18.410 "cntlid": 103, 00:17:18.410 "qid": 0, 00:17:18.410 "state": "enabled", 00:17:18.410 "thread": "nvmf_tgt_poll_group_000", 00:17:18.410 "listen_address": { 00:17:18.410 "trtype": "TCP", 00:17:18.410 "adrfam": "IPv4", 00:17:18.410 "traddr": "10.0.0.2", 00:17:18.410 "trsvcid": "4420" 00:17:18.410 }, 00:17:18.410 "peer_address": { 00:17:18.410 "trtype": "TCP", 00:17:18.410 "adrfam": "IPv4", 00:17:18.410 "traddr": "10.0.0.1", 00:17:18.410 "trsvcid": "59636" 00:17:18.410 }, 00:17:18.410 "auth": { 00:17:18.410 "state": "completed", 00:17:18.410 "digest": "sha512", 00:17:18.410 "dhgroup": "null" 00:17:18.410 } 00:17:18.410 } 00:17:18.410 ]' 00:17:18.410 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:18.410 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:18.410 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:18.410 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ null == \n\u\l\l ]] 00:17:18.410 18:03:11 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:18.410 18:03:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:18.410 18:03:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:18.410 18:03:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:18.669 18:03:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:03:MDkyZTk1N2U3YjVlMDJjNThmOTAwMDMzY2YzZWQ3ZjZiYTk2ZDA3MzczYTM5MTgyNDYxMjY5ZjBjMDlhNGMzZRaivds=: 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:19.236 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@92 -- # for dhgroup in "${dhgroups[@]}" 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 ffdhe2048 0 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe2048 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key0 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:19.236 18:03:12 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:19.515 00:17:19.515 18:03:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:19.515 18:03:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:19.515 18:03:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:19.772 18:03:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:19.772 18:03:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:19.772 18:03:13 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:19.772 18:03:13 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:19.772 18:03:13 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:19.772 18:03:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:19.772 { 00:17:19.772 "cntlid": 105, 00:17:19.772 "qid": 0, 00:17:19.772 "state": "enabled", 00:17:19.773 "thread": "nvmf_tgt_poll_group_000", 00:17:19.773 "listen_address": { 00:17:19.773 "trtype": "TCP", 00:17:19.773 "adrfam": "IPv4", 00:17:19.773 "traddr": "10.0.0.2", 00:17:19.773 "trsvcid": "4420" 00:17:19.773 }, 00:17:19.773 "peer_address": { 00:17:19.773 "trtype": "TCP", 00:17:19.773 "adrfam": "IPv4", 00:17:19.773 "traddr": "10.0.0.1", 00:17:19.773 "trsvcid": "59658" 00:17:19.773 }, 00:17:19.773 "auth": { 00:17:19.773 "state": "completed", 00:17:19.773 "digest": "sha512", 00:17:19.773 "dhgroup": "ffdhe2048" 00:17:19.773 } 00:17:19.773 } 00:17:19.773 ]' 00:17:19.773 18:03:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:19.773 18:03:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:19.773 18:03:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:19.773 18:03:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:17:19.773 18:03:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:19.773 18:03:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:19.773 18:03:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:19.773 18:03:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:20.031 18:03:13 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:00:ZDc0NTQyMWE2OWE3NzMyMzhhZTE1NDNiYzY0MmRlNGJhY2Q1NjlmNTk1NGFlN2VlXS98/Q==: --dhchap-ctrl-secret DHHC-1:03:NmM0ODExMWYzNDJlOTA1ZjgyNTliZThhZmQ2NGY1YTMwZGFmYmVkMTY3MmFiNjRkZDdiZTAwNmRhZDYxYmI1N4Bp9x8=: 00:17:20.598 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:20.598 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:20.598 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:20.598 18:03:14 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:20.598 18:03:14 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:20.599 18:03:14 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:20.599 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:20.599 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:17:20.599 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:17:20.857 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 ffdhe2048 1 00:17:20.857 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:20.857 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:20.857 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe2048 00:17:20.858 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key1 00:17:20.858 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:20.858 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:20.858 18:03:14 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:20.858 18:03:14 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:20.858 18:03:14 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:20.858 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:20.858 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:21.117 00:17:21.117 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:21.117 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:21.117 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:21.117 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:21.117 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:21.117 18:03:14 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:21.117 18:03:14 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:21.117 18:03:14 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:21.117 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:21.117 { 00:17:21.117 "cntlid": 107, 00:17:21.117 "qid": 0, 00:17:21.117 "state": "enabled", 00:17:21.117 "thread": "nvmf_tgt_poll_group_000", 00:17:21.117 "listen_address": { 00:17:21.117 "trtype": "TCP", 00:17:21.117 "adrfam": "IPv4", 00:17:21.117 "traddr": "10.0.0.2", 00:17:21.117 "trsvcid": "4420" 00:17:21.117 }, 00:17:21.117 "peer_address": { 00:17:21.117 "trtype": "TCP", 00:17:21.117 "adrfam": "IPv4", 00:17:21.117 "traddr": "10.0.0.1", 00:17:21.117 "trsvcid": "59682" 00:17:21.117 }, 00:17:21.117 "auth": { 00:17:21.117 "state": "completed", 00:17:21.117 "digest": "sha512", 00:17:21.117 "dhgroup": "ffdhe2048" 00:17:21.117 } 00:17:21.117 } 00:17:21.117 ]' 00:17:21.117 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:21.376 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:21.376 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:21.376 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:17:21.376 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:21.376 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:21.376 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:21.376 18:03:14 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:21.634 18:03:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:01:ZDMzNjZhODE5ZDkwZjc1YzJlMWRhMzJmNzNkZWZjM2TGg2/J: --dhchap-ctrl-secret DHHC-1:02:OWFmZjc0MDk4ZDJmZWMyYWFjNmJmNmM1MzZkMjUwMDZlY2RmOGMxNjQxNzZkZDdloY1p5Q==: 00:17:22.201 18:03:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:22.201 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:22.201 18:03:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:22.201 18:03:15 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:22.201 18:03:15 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:22.201 18:03:15 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:22.201 18:03:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:22.201 18:03:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:17:22.201 18:03:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:17:22.201 18:03:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 ffdhe2048 2 00:17:22.201 18:03:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:22.201 18:03:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:22.201 18:03:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe2048 00:17:22.201 18:03:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key2 00:17:22.201 18:03:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:22.201 18:03:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:22.201 18:03:15 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:22.201 18:03:15 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:22.201 18:03:15 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:22.202 18:03:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:22.202 18:03:15 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:22.460 00:17:22.460 18:03:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:22.460 18:03:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:22.460 18:03:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:22.719 18:03:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:22.719 18:03:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:22.719 18:03:16 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:22.719 18:03:16 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:22.719 18:03:16 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:22.719 18:03:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:22.719 { 00:17:22.719 "cntlid": 109, 00:17:22.719 "qid": 0, 00:17:22.719 "state": "enabled", 00:17:22.719 "thread": "nvmf_tgt_poll_group_000", 00:17:22.719 "listen_address": { 00:17:22.719 "trtype": "TCP", 00:17:22.719 "adrfam": "IPv4", 00:17:22.719 "traddr": "10.0.0.2", 00:17:22.719 "trsvcid": "4420" 00:17:22.719 }, 00:17:22.719 "peer_address": { 00:17:22.719 "trtype": "TCP", 00:17:22.719 "adrfam": "IPv4", 00:17:22.719 "traddr": "10.0.0.1", 00:17:22.719 "trsvcid": "59708" 00:17:22.719 }, 00:17:22.719 "auth": { 00:17:22.719 "state": "completed", 00:17:22.719 "digest": "sha512", 00:17:22.719 "dhgroup": "ffdhe2048" 00:17:22.719 } 00:17:22.719 } 00:17:22.719 ]' 00:17:22.719 18:03:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:22.719 18:03:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:22.719 18:03:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:22.719 18:03:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:17:22.719 18:03:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:22.719 18:03:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:22.719 18:03:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:22.719 18:03:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:22.978 18:03:16 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:02:M2JjZDM2NWNhZmQ5MTRjN2UwY2Q2YjhiOWUwYmE1MzZiNzNiZDYwNjZhZGJkYjM3IIKM4Q==: --dhchap-ctrl-secret DHHC-1:01:OTFlOTkyN2M2NzZiZWZkNWRjOTE4ZmUxNTUyYjI3ODS55lrp: 00:17:23.543 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:23.543 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:23.543 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:23.543 18:03:17 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:23.543 18:03:17 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:23.543 18:03:17 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:23.543 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:23.543 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:17:23.543 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:17:23.802 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 ffdhe2048 3 00:17:23.802 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:23.802 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:23.802 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe2048 00:17:23.802 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key3 00:17:23.802 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:23.802 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key3 00:17:23.802 18:03:17 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:23.802 18:03:17 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:23.802 18:03:17 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:23.802 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:23.802 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:24.060 00:17:24.060 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:24.060 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:24.061 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:24.061 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:24.061 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:24.061 18:03:17 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:24.061 18:03:17 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:24.319 18:03:17 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:24.319 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:24.319 { 00:17:24.319 "cntlid": 111, 00:17:24.319 "qid": 0, 00:17:24.319 "state": "enabled", 00:17:24.319 "thread": "nvmf_tgt_poll_group_000", 00:17:24.319 "listen_address": { 00:17:24.319 "trtype": "TCP", 00:17:24.319 "adrfam": "IPv4", 00:17:24.319 "traddr": "10.0.0.2", 00:17:24.319 "trsvcid": "4420" 00:17:24.319 }, 00:17:24.319 "peer_address": { 00:17:24.319 "trtype": "TCP", 00:17:24.319 "adrfam": "IPv4", 00:17:24.319 "traddr": "10.0.0.1", 00:17:24.319 "trsvcid": "48112" 00:17:24.319 }, 00:17:24.319 "auth": { 00:17:24.319 "state": "completed", 00:17:24.319 "digest": "sha512", 00:17:24.319 "dhgroup": "ffdhe2048" 00:17:24.319 } 00:17:24.319 } 00:17:24.319 ]' 00:17:24.319 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:24.319 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:24.320 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:24.320 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe2048 == \f\f\d\h\e\2\0\4\8 ]] 00:17:24.320 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:24.320 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:24.320 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:24.320 18:03:17 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:24.578 18:03:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:03:MDkyZTk1N2U3YjVlMDJjNThmOTAwMDMzY2YzZWQ3ZjZiYTk2ZDA3MzczYTM5MTgyNDYxMjY5ZjBjMDlhNGMzZRaivds=: 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:25.146 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@92 -- # for dhgroup in "${dhgroups[@]}" 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 ffdhe3072 0 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe3072 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key0 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:25.146 18:03:18 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:25.405 00:17:25.405 18:03:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:25.405 18:03:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:25.405 18:03:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:25.664 18:03:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:25.664 18:03:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:25.664 18:03:19 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:25.664 18:03:19 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:25.664 18:03:19 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:25.664 18:03:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:25.664 { 00:17:25.664 "cntlid": 113, 00:17:25.664 "qid": 0, 00:17:25.664 "state": "enabled", 00:17:25.664 "thread": "nvmf_tgt_poll_group_000", 00:17:25.664 "listen_address": { 00:17:25.664 "trtype": "TCP", 00:17:25.664 "adrfam": "IPv4", 00:17:25.664 "traddr": "10.0.0.2", 00:17:25.664 "trsvcid": "4420" 00:17:25.664 }, 00:17:25.664 "peer_address": { 00:17:25.664 "trtype": "TCP", 00:17:25.664 "adrfam": "IPv4", 00:17:25.664 "traddr": "10.0.0.1", 00:17:25.664 "trsvcid": "48136" 00:17:25.664 }, 00:17:25.664 "auth": { 00:17:25.664 "state": "completed", 00:17:25.664 "digest": "sha512", 00:17:25.664 "dhgroup": "ffdhe3072" 00:17:25.664 } 00:17:25.664 } 00:17:25.664 ]' 00:17:25.664 18:03:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:25.664 18:03:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:25.664 18:03:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:25.664 18:03:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:17:25.664 18:03:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:25.923 18:03:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:25.923 18:03:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:25.923 18:03:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:25.923 18:03:19 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:00:ZDc0NTQyMWE2OWE3NzMyMzhhZTE1NDNiYzY0MmRlNGJhY2Q1NjlmNTk1NGFlN2VlXS98/Q==: --dhchap-ctrl-secret DHHC-1:03:NmM0ODExMWYzNDJlOTA1ZjgyNTliZThhZmQ2NGY1YTMwZGFmYmVkMTY3MmFiNjRkZDdiZTAwNmRhZDYxYmI1N4Bp9x8=: 00:17:26.491 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:26.491 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:26.491 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:26.491 18:03:20 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:26.491 18:03:20 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:26.491 18:03:20 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:26.491 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:26.491 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:17:26.491 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:17:26.750 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 ffdhe3072 1 00:17:26.750 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:26.750 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:26.750 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe3072 00:17:26.750 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key1 00:17:26.750 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:26.750 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:26.750 18:03:20 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:26.750 18:03:20 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:26.750 18:03:20 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:26.750 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:26.750 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:27.008 00:17:27.008 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:27.008 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:27.008 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:27.267 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:27.267 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:27.267 18:03:20 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:27.267 18:03:20 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:27.267 18:03:20 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:27.267 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:27.267 { 00:17:27.267 "cntlid": 115, 00:17:27.267 "qid": 0, 00:17:27.267 "state": "enabled", 00:17:27.267 "thread": "nvmf_tgt_poll_group_000", 00:17:27.267 "listen_address": { 00:17:27.267 "trtype": "TCP", 00:17:27.267 "adrfam": "IPv4", 00:17:27.267 "traddr": "10.0.0.2", 00:17:27.267 "trsvcid": "4420" 00:17:27.267 }, 00:17:27.267 "peer_address": { 00:17:27.267 "trtype": "TCP", 00:17:27.267 "adrfam": "IPv4", 00:17:27.267 "traddr": "10.0.0.1", 00:17:27.267 "trsvcid": "48164" 00:17:27.267 }, 00:17:27.267 "auth": { 00:17:27.267 "state": "completed", 00:17:27.267 "digest": "sha512", 00:17:27.267 "dhgroup": "ffdhe3072" 00:17:27.267 } 00:17:27.267 } 00:17:27.267 ]' 00:17:27.267 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:27.267 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:27.267 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:27.267 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:17:27.267 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:27.267 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:27.267 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:27.267 18:03:20 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:27.526 18:03:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:01:ZDMzNjZhODE5ZDkwZjc1YzJlMWRhMzJmNzNkZWZjM2TGg2/J: --dhchap-ctrl-secret DHHC-1:02:OWFmZjc0MDk4ZDJmZWMyYWFjNmJmNmM1MzZkMjUwMDZlY2RmOGMxNjQxNzZkZDdloY1p5Q==: 00:17:28.093 18:03:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:28.093 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:28.093 18:03:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:28.093 18:03:21 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:28.093 18:03:21 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:28.093 18:03:21 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:28.093 18:03:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:28.093 18:03:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:17:28.093 18:03:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:17:28.352 18:03:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 ffdhe3072 2 00:17:28.352 18:03:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:28.352 18:03:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:28.352 18:03:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe3072 00:17:28.352 18:03:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key2 00:17:28.352 18:03:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:28.352 18:03:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:28.352 18:03:21 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:28.352 18:03:21 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:28.352 18:03:21 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:28.352 18:03:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:28.352 18:03:21 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:28.611 00:17:28.611 18:03:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:28.611 18:03:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:28.611 18:03:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:28.611 18:03:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:28.611 18:03:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:28.611 18:03:22 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:28.611 18:03:22 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:28.871 18:03:22 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:28.871 18:03:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:28.871 { 00:17:28.871 "cntlid": 117, 00:17:28.871 "qid": 0, 00:17:28.871 "state": "enabled", 00:17:28.871 "thread": "nvmf_tgt_poll_group_000", 00:17:28.871 "listen_address": { 00:17:28.871 "trtype": "TCP", 00:17:28.871 "adrfam": "IPv4", 00:17:28.871 "traddr": "10.0.0.2", 00:17:28.871 "trsvcid": "4420" 00:17:28.871 }, 00:17:28.871 "peer_address": { 00:17:28.871 "trtype": "TCP", 00:17:28.871 "adrfam": "IPv4", 00:17:28.871 "traddr": "10.0.0.1", 00:17:28.871 "trsvcid": "48198" 00:17:28.871 }, 00:17:28.871 "auth": { 00:17:28.871 "state": "completed", 00:17:28.871 "digest": "sha512", 00:17:28.871 "dhgroup": "ffdhe3072" 00:17:28.871 } 00:17:28.871 } 00:17:28.871 ]' 00:17:28.871 18:03:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:28.871 18:03:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:28.871 18:03:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:28.871 18:03:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:17:28.871 18:03:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:28.871 18:03:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:28.871 18:03:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:28.871 18:03:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:29.130 18:03:22 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:02:M2JjZDM2NWNhZmQ5MTRjN2UwY2Q2YjhiOWUwYmE1MzZiNzNiZDYwNjZhZGJkYjM3IIKM4Q==: --dhchap-ctrl-secret DHHC-1:01:OTFlOTkyN2M2NzZiZWZkNWRjOTE4ZmUxNTUyYjI3ODS55lrp: 00:17:29.719 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:29.719 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:29.719 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:29.719 18:03:23 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:29.719 18:03:23 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:29.719 18:03:23 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:29.719 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:29.719 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:17:29.719 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:17:29.719 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 ffdhe3072 3 00:17:29.719 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:29.719 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:29.719 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe3072 00:17:29.719 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key3 00:17:29.719 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:29.719 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key3 00:17:29.719 18:03:23 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:29.719 18:03:23 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:29.977 18:03:23 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:29.977 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:29.977 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:29.977 00:17:29.977 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:29.977 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:29.977 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:30.237 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:30.237 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:30.237 18:03:23 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:30.237 18:03:23 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:30.237 18:03:23 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:30.237 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:30.237 { 00:17:30.237 "cntlid": 119, 00:17:30.237 "qid": 0, 00:17:30.237 "state": "enabled", 00:17:30.237 "thread": "nvmf_tgt_poll_group_000", 00:17:30.237 "listen_address": { 00:17:30.237 "trtype": "TCP", 00:17:30.237 "adrfam": "IPv4", 00:17:30.237 "traddr": "10.0.0.2", 00:17:30.237 "trsvcid": "4420" 00:17:30.237 }, 00:17:30.237 "peer_address": { 00:17:30.237 "trtype": "TCP", 00:17:30.237 "adrfam": "IPv4", 00:17:30.237 "traddr": "10.0.0.1", 00:17:30.237 "trsvcid": "48226" 00:17:30.237 }, 00:17:30.237 "auth": { 00:17:30.237 "state": "completed", 00:17:30.237 "digest": "sha512", 00:17:30.237 "dhgroup": "ffdhe3072" 00:17:30.237 } 00:17:30.237 } 00:17:30.237 ]' 00:17:30.237 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:30.237 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:30.237 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:30.237 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe3072 == \f\f\d\h\e\3\0\7\2 ]] 00:17:30.237 18:03:23 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:30.496 18:03:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:30.496 18:03:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:30.496 18:03:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:30.496 18:03:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:03:MDkyZTk1N2U3YjVlMDJjNThmOTAwMDMzY2YzZWQ3ZjZiYTk2ZDA3MzczYTM5MTgyNDYxMjY5ZjBjMDlhNGMzZRaivds=: 00:17:31.064 18:03:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:31.064 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:31.064 18:03:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:31.064 18:03:24 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:31.064 18:03:24 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:31.064 18:03:24 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:31.064 18:03:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@92 -- # for dhgroup in "${dhgroups[@]}" 00:17:31.064 18:03:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:31.064 18:03:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:17:31.064 18:03:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:17:31.323 18:03:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 ffdhe4096 0 00:17:31.323 18:03:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:31.323 18:03:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:31.323 18:03:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe4096 00:17:31.323 18:03:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key0 00:17:31.323 18:03:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:31.323 18:03:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:31.323 18:03:24 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:31.323 18:03:24 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:31.323 18:03:24 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:31.323 18:03:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:31.323 18:03:24 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:31.607 00:17:31.607 18:03:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:31.607 18:03:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:31.607 18:03:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:31.911 18:03:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:31.911 18:03:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:31.911 18:03:25 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:31.911 18:03:25 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:31.911 18:03:25 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:31.911 18:03:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:31.911 { 00:17:31.911 "cntlid": 121, 00:17:31.911 "qid": 0, 00:17:31.911 "state": "enabled", 00:17:31.911 "thread": "nvmf_tgt_poll_group_000", 00:17:31.911 "listen_address": { 00:17:31.911 "trtype": "TCP", 00:17:31.911 "adrfam": "IPv4", 00:17:31.911 "traddr": "10.0.0.2", 00:17:31.911 "trsvcid": "4420" 00:17:31.911 }, 00:17:31.911 "peer_address": { 00:17:31.911 "trtype": "TCP", 00:17:31.911 "adrfam": "IPv4", 00:17:31.911 "traddr": "10.0.0.1", 00:17:31.911 "trsvcid": "48254" 00:17:31.911 }, 00:17:31.911 "auth": { 00:17:31.911 "state": "completed", 00:17:31.911 "digest": "sha512", 00:17:31.911 "dhgroup": "ffdhe4096" 00:17:31.911 } 00:17:31.911 } 00:17:31.911 ]' 00:17:31.911 18:03:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:31.911 18:03:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:31.911 18:03:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:31.911 18:03:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:17:31.911 18:03:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:31.911 18:03:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:31.911 18:03:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:31.911 18:03:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:32.169 18:03:25 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:00:ZDc0NTQyMWE2OWE3NzMyMzhhZTE1NDNiYzY0MmRlNGJhY2Q1NjlmNTk1NGFlN2VlXS98/Q==: --dhchap-ctrl-secret DHHC-1:03:NmM0ODExMWYzNDJlOTA1ZjgyNTliZThhZmQ2NGY1YTMwZGFmYmVkMTY3MmFiNjRkZDdiZTAwNmRhZDYxYmI1N4Bp9x8=: 00:17:32.736 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:32.736 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:32.736 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:32.736 18:03:26 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:32.736 18:03:26 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:32.736 18:03:26 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:32.736 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:32.736 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:17:32.736 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:17:32.995 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 ffdhe4096 1 00:17:32.995 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:32.995 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:32.995 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe4096 00:17:32.995 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key1 00:17:32.995 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:32.995 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:32.995 18:03:26 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:32.995 18:03:26 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:32.995 18:03:26 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:32.995 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:32.995 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:33.254 00:17:33.254 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:33.254 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:33.254 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:33.254 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:33.254 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:33.254 18:03:26 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:33.254 18:03:26 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:33.254 18:03:26 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:33.254 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:33.254 { 00:17:33.254 "cntlid": 123, 00:17:33.254 "qid": 0, 00:17:33.254 "state": "enabled", 00:17:33.254 "thread": "nvmf_tgt_poll_group_000", 00:17:33.254 "listen_address": { 00:17:33.254 "trtype": "TCP", 00:17:33.254 "adrfam": "IPv4", 00:17:33.254 "traddr": "10.0.0.2", 00:17:33.254 "trsvcid": "4420" 00:17:33.254 }, 00:17:33.254 "peer_address": { 00:17:33.254 "trtype": "TCP", 00:17:33.254 "adrfam": "IPv4", 00:17:33.254 "traddr": "10.0.0.1", 00:17:33.254 "trsvcid": "48268" 00:17:33.254 }, 00:17:33.254 "auth": { 00:17:33.254 "state": "completed", 00:17:33.254 "digest": "sha512", 00:17:33.254 "dhgroup": "ffdhe4096" 00:17:33.254 } 00:17:33.254 } 00:17:33.254 ]' 00:17:33.254 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:33.254 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:33.254 18:03:26 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:33.514 18:03:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:17:33.514 18:03:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:33.514 18:03:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:33.514 18:03:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:33.514 18:03:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:33.514 18:03:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:01:ZDMzNjZhODE5ZDkwZjc1YzJlMWRhMzJmNzNkZWZjM2TGg2/J: --dhchap-ctrl-secret DHHC-1:02:OWFmZjc0MDk4ZDJmZWMyYWFjNmJmNmM1MzZkMjUwMDZlY2RmOGMxNjQxNzZkZDdloY1p5Q==: 00:17:34.082 18:03:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:34.342 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:34.342 18:03:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:34.342 18:03:27 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:34.342 18:03:27 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:34.342 18:03:27 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:34.342 18:03:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:34.342 18:03:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:17:34.342 18:03:27 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:17:34.342 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 ffdhe4096 2 00:17:34.342 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:34.342 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:34.342 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe4096 00:17:34.342 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key2 00:17:34.342 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:34.342 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:34.342 18:03:28 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:34.342 18:03:28 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:34.342 18:03:28 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:34.342 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:34.342 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:34.608 00:17:34.608 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:34.608 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:34.608 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:34.868 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:34.868 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:34.868 18:03:28 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:34.868 18:03:28 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:34.868 18:03:28 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:34.868 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:34.868 { 00:17:34.868 "cntlid": 125, 00:17:34.868 "qid": 0, 00:17:34.868 "state": "enabled", 00:17:34.868 "thread": "nvmf_tgt_poll_group_000", 00:17:34.868 "listen_address": { 00:17:34.868 "trtype": "TCP", 00:17:34.868 "adrfam": "IPv4", 00:17:34.868 "traddr": "10.0.0.2", 00:17:34.868 "trsvcid": "4420" 00:17:34.868 }, 00:17:34.868 "peer_address": { 00:17:34.868 "trtype": "TCP", 00:17:34.868 "adrfam": "IPv4", 00:17:34.868 "traddr": "10.0.0.1", 00:17:34.868 "trsvcid": "52592" 00:17:34.868 }, 00:17:34.868 "auth": { 00:17:34.868 "state": "completed", 00:17:34.868 "digest": "sha512", 00:17:34.868 "dhgroup": "ffdhe4096" 00:17:34.868 } 00:17:34.868 } 00:17:34.868 ]' 00:17:34.868 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:34.868 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:34.868 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:34.868 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:17:34.868 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:34.868 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:34.868 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:34.868 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:35.127 18:03:28 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:02:M2JjZDM2NWNhZmQ5MTRjN2UwY2Q2YjhiOWUwYmE1MzZiNzNiZDYwNjZhZGJkYjM3IIKM4Q==: --dhchap-ctrl-secret DHHC-1:01:OTFlOTkyN2M2NzZiZWZkNWRjOTE4ZmUxNTUyYjI3ODS55lrp: 00:17:35.695 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:35.695 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:35.695 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:35.695 18:03:29 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:35.695 18:03:29 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:35.695 18:03:29 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:35.695 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:35.695 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:17:35.695 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:17:35.955 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 ffdhe4096 3 00:17:35.955 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:35.955 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:35.955 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe4096 00:17:35.955 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key3 00:17:35.955 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:35.955 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key3 00:17:35.955 18:03:29 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:35.955 18:03:29 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:35.955 18:03:29 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:35.955 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:35.955 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:36.213 00:17:36.213 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:36.213 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:36.213 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:36.472 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:36.472 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:36.472 18:03:29 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:36.472 18:03:29 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:36.472 18:03:29 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:36.472 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:36.472 { 00:17:36.472 "cntlid": 127, 00:17:36.472 "qid": 0, 00:17:36.472 "state": "enabled", 00:17:36.472 "thread": "nvmf_tgt_poll_group_000", 00:17:36.472 "listen_address": { 00:17:36.472 "trtype": "TCP", 00:17:36.472 "adrfam": "IPv4", 00:17:36.472 "traddr": "10.0.0.2", 00:17:36.472 "trsvcid": "4420" 00:17:36.472 }, 00:17:36.472 "peer_address": { 00:17:36.472 "trtype": "TCP", 00:17:36.472 "adrfam": "IPv4", 00:17:36.472 "traddr": "10.0.0.1", 00:17:36.472 "trsvcid": "52622" 00:17:36.472 }, 00:17:36.472 "auth": { 00:17:36.472 "state": "completed", 00:17:36.472 "digest": "sha512", 00:17:36.472 "dhgroup": "ffdhe4096" 00:17:36.472 } 00:17:36.472 } 00:17:36.472 ]' 00:17:36.472 18:03:29 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:36.472 18:03:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:36.472 18:03:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:36.472 18:03:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe4096 == \f\f\d\h\e\4\0\9\6 ]] 00:17:36.472 18:03:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:36.472 18:03:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:36.472 18:03:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:36.472 18:03:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:36.731 18:03:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:03:MDkyZTk1N2U3YjVlMDJjNThmOTAwMDMzY2YzZWQ3ZjZiYTk2ZDA3MzczYTM5MTgyNDYxMjY5ZjBjMDlhNGMzZRaivds=: 00:17:37.298 18:03:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:37.298 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:37.298 18:03:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:37.298 18:03:30 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:37.298 18:03:30 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:37.298 18:03:30 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:37.298 18:03:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@92 -- # for dhgroup in "${dhgroups[@]}" 00:17:37.298 18:03:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:37.298 18:03:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:17:37.298 18:03:30 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:17:37.557 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 ffdhe6144 0 00:17:37.557 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:37.557 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:37.557 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe6144 00:17:37.557 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key0 00:17:37.557 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:37.557 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:37.557 18:03:31 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:37.557 18:03:31 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:37.557 18:03:31 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:37.557 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:37.557 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:37.816 00:17:37.816 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:37.816 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:37.816 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:38.075 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:38.075 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:38.075 18:03:31 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:38.075 18:03:31 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:38.075 18:03:31 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:38.075 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:38.075 { 00:17:38.075 "cntlid": 129, 00:17:38.075 "qid": 0, 00:17:38.075 "state": "enabled", 00:17:38.075 "thread": "nvmf_tgt_poll_group_000", 00:17:38.075 "listen_address": { 00:17:38.075 "trtype": "TCP", 00:17:38.075 "adrfam": "IPv4", 00:17:38.075 "traddr": "10.0.0.2", 00:17:38.075 "trsvcid": "4420" 00:17:38.075 }, 00:17:38.075 "peer_address": { 00:17:38.075 "trtype": "TCP", 00:17:38.075 "adrfam": "IPv4", 00:17:38.075 "traddr": "10.0.0.1", 00:17:38.075 "trsvcid": "52638" 00:17:38.075 }, 00:17:38.075 "auth": { 00:17:38.075 "state": "completed", 00:17:38.075 "digest": "sha512", 00:17:38.075 "dhgroup": "ffdhe6144" 00:17:38.075 } 00:17:38.075 } 00:17:38.075 ]' 00:17:38.075 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:38.075 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:38.075 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:38.075 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:17:38.075 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:38.075 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:38.075 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:38.075 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:38.334 18:03:31 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:00:ZDc0NTQyMWE2OWE3NzMyMzhhZTE1NDNiYzY0MmRlNGJhY2Q1NjlmNTk1NGFlN2VlXS98/Q==: --dhchap-ctrl-secret DHHC-1:03:NmM0ODExMWYzNDJlOTA1ZjgyNTliZThhZmQ2NGY1YTMwZGFmYmVkMTY3MmFiNjRkZDdiZTAwNmRhZDYxYmI1N4Bp9x8=: 00:17:38.901 18:03:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:38.901 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:38.901 18:03:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:38.901 18:03:32 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:38.901 18:03:32 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:38.901 18:03:32 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:38.901 18:03:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:38.901 18:03:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:17:38.901 18:03:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:17:39.159 18:03:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 ffdhe6144 1 00:17:39.159 18:03:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:39.159 18:03:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:39.159 18:03:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe6144 00:17:39.159 18:03:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key1 00:17:39.159 18:03:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:39.159 18:03:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:39.159 18:03:32 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:39.159 18:03:32 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:39.159 18:03:32 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:39.159 18:03:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:39.159 18:03:32 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:39.418 00:17:39.418 18:03:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:39.418 18:03:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:39.418 18:03:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:39.688 18:03:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:39.688 18:03:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:39.688 18:03:33 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:39.688 18:03:33 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:39.688 18:03:33 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:39.688 18:03:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:39.688 { 00:17:39.688 "cntlid": 131, 00:17:39.688 "qid": 0, 00:17:39.688 "state": "enabled", 00:17:39.688 "thread": "nvmf_tgt_poll_group_000", 00:17:39.688 "listen_address": { 00:17:39.688 "trtype": "TCP", 00:17:39.688 "adrfam": "IPv4", 00:17:39.688 "traddr": "10.0.0.2", 00:17:39.688 "trsvcid": "4420" 00:17:39.688 }, 00:17:39.688 "peer_address": { 00:17:39.688 "trtype": "TCP", 00:17:39.688 "adrfam": "IPv4", 00:17:39.688 "traddr": "10.0.0.1", 00:17:39.688 "trsvcid": "52660" 00:17:39.688 }, 00:17:39.688 "auth": { 00:17:39.689 "state": "completed", 00:17:39.689 "digest": "sha512", 00:17:39.689 "dhgroup": "ffdhe6144" 00:17:39.689 } 00:17:39.689 } 00:17:39.689 ]' 00:17:39.689 18:03:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:39.689 18:03:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:39.689 18:03:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:39.689 18:03:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:17:39.689 18:03:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:39.689 18:03:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:39.689 18:03:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:39.689 18:03:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:39.955 18:03:33 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:01:ZDMzNjZhODE5ZDkwZjc1YzJlMWRhMzJmNzNkZWZjM2TGg2/J: --dhchap-ctrl-secret DHHC-1:02:OWFmZjc0MDk4ZDJmZWMyYWFjNmJmNmM1MzZkMjUwMDZlY2RmOGMxNjQxNzZkZDdloY1p5Q==: 00:17:40.522 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:40.522 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:40.522 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:40.522 18:03:34 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:40.522 18:03:34 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:40.522 18:03:34 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:40.522 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:40.523 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:17:40.523 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:17:40.782 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 ffdhe6144 2 00:17:40.782 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:40.782 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:40.782 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe6144 00:17:40.782 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key2 00:17:40.782 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:40.782 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:40.782 18:03:34 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:40.782 18:03:34 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:40.782 18:03:34 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:40.782 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:40.782 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:41.040 00:17:41.040 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:41.040 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:41.040 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:41.299 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:41.299 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:41.299 18:03:34 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:41.299 18:03:34 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:41.299 18:03:34 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:41.299 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:41.299 { 00:17:41.299 "cntlid": 133, 00:17:41.299 "qid": 0, 00:17:41.299 "state": "enabled", 00:17:41.299 "thread": "nvmf_tgt_poll_group_000", 00:17:41.299 "listen_address": { 00:17:41.299 "trtype": "TCP", 00:17:41.299 "adrfam": "IPv4", 00:17:41.299 "traddr": "10.0.0.2", 00:17:41.299 "trsvcid": "4420" 00:17:41.299 }, 00:17:41.299 "peer_address": { 00:17:41.299 "trtype": "TCP", 00:17:41.299 "adrfam": "IPv4", 00:17:41.299 "traddr": "10.0.0.1", 00:17:41.299 "trsvcid": "52692" 00:17:41.299 }, 00:17:41.299 "auth": { 00:17:41.299 "state": "completed", 00:17:41.299 "digest": "sha512", 00:17:41.299 "dhgroup": "ffdhe6144" 00:17:41.299 } 00:17:41.299 } 00:17:41.299 ]' 00:17:41.299 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:41.299 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:41.299 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:41.299 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:17:41.299 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:41.299 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:41.299 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:41.299 18:03:34 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:41.557 18:03:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:02:M2JjZDM2NWNhZmQ5MTRjN2UwY2Q2YjhiOWUwYmE1MzZiNzNiZDYwNjZhZGJkYjM3IIKM4Q==: --dhchap-ctrl-secret DHHC-1:01:OTFlOTkyN2M2NzZiZWZkNWRjOTE4ZmUxNTUyYjI3ODS55lrp: 00:17:42.124 18:03:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:42.124 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:42.124 18:03:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:42.124 18:03:35 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:42.124 18:03:35 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:42.124 18:03:35 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:42.124 18:03:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:42.124 18:03:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:17:42.124 18:03:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:17:42.384 18:03:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 ffdhe6144 3 00:17:42.384 18:03:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:42.384 18:03:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:42.384 18:03:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe6144 00:17:42.384 18:03:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key3 00:17:42.384 18:03:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:42.384 18:03:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key3 00:17:42.384 18:03:35 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:42.384 18:03:35 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:42.384 18:03:35 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:42.384 18:03:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:42.384 18:03:35 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:42.643 00:17:42.643 18:03:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:42.643 18:03:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:42.643 18:03:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:42.902 18:03:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:42.902 18:03:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:42.902 18:03:36 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:42.902 18:03:36 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:42.902 18:03:36 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:42.902 18:03:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:42.902 { 00:17:42.902 "cntlid": 135, 00:17:42.902 "qid": 0, 00:17:42.902 "state": "enabled", 00:17:42.902 "thread": "nvmf_tgt_poll_group_000", 00:17:42.902 "listen_address": { 00:17:42.902 "trtype": "TCP", 00:17:42.902 "adrfam": "IPv4", 00:17:42.902 "traddr": "10.0.0.2", 00:17:42.902 "trsvcid": "4420" 00:17:42.902 }, 00:17:42.902 "peer_address": { 00:17:42.902 "trtype": "TCP", 00:17:42.902 "adrfam": "IPv4", 00:17:42.902 "traddr": "10.0.0.1", 00:17:42.902 "trsvcid": "52722" 00:17:42.902 }, 00:17:42.902 "auth": { 00:17:42.902 "state": "completed", 00:17:42.902 "digest": "sha512", 00:17:42.902 "dhgroup": "ffdhe6144" 00:17:42.902 } 00:17:42.902 } 00:17:42.902 ]' 00:17:42.902 18:03:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:42.902 18:03:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:42.902 18:03:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:42.902 18:03:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe6144 == \f\f\d\h\e\6\1\4\4 ]] 00:17:42.902 18:03:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:42.902 18:03:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:42.903 18:03:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:42.903 18:03:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:43.161 18:03:36 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:03:MDkyZTk1N2U3YjVlMDJjNThmOTAwMDMzY2YzZWQ3ZjZiYTk2ZDA3MzczYTM5MTgyNDYxMjY5ZjBjMDlhNGMzZRaivds=: 00:17:43.729 18:03:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:43.729 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:43.729 18:03:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:43.729 18:03:37 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:43.729 18:03:37 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:43.729 18:03:37 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:43.729 18:03:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@92 -- # for dhgroup in "${dhgroups[@]}" 00:17:43.729 18:03:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:43.729 18:03:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:17:43.729 18:03:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:17:43.729 18:03:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 ffdhe8192 0 00:17:43.729 18:03:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:43.729 18:03:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:43.729 18:03:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe8192 00:17:43.729 18:03:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key0 00:17:43.729 18:03:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:43.729 18:03:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:43.729 18:03:37 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:43.729 18:03:37 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:43.988 18:03:37 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:43.988 18:03:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:43.988 18:03:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:44.247 00:17:44.247 18:03:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:44.247 18:03:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:44.247 18:03:37 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:44.507 18:03:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:44.507 18:03:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:44.507 18:03:38 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:44.507 18:03:38 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:44.507 18:03:38 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:44.507 18:03:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:44.507 { 00:17:44.507 "cntlid": 137, 00:17:44.507 "qid": 0, 00:17:44.507 "state": "enabled", 00:17:44.507 "thread": "nvmf_tgt_poll_group_000", 00:17:44.507 "listen_address": { 00:17:44.507 "trtype": "TCP", 00:17:44.507 "adrfam": "IPv4", 00:17:44.507 "traddr": "10.0.0.2", 00:17:44.507 "trsvcid": "4420" 00:17:44.507 }, 00:17:44.507 "peer_address": { 00:17:44.507 "trtype": "TCP", 00:17:44.507 "adrfam": "IPv4", 00:17:44.507 "traddr": "10.0.0.1", 00:17:44.507 "trsvcid": "59974" 00:17:44.507 }, 00:17:44.507 "auth": { 00:17:44.507 "state": "completed", 00:17:44.507 "digest": "sha512", 00:17:44.507 "dhgroup": "ffdhe8192" 00:17:44.507 } 00:17:44.507 } 00:17:44.507 ]' 00:17:44.507 18:03:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:44.507 18:03:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:44.507 18:03:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:44.507 18:03:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:44.507 18:03:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:44.507 18:03:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:44.507 18:03:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:44.767 18:03:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:44.767 18:03:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:00:ZDc0NTQyMWE2OWE3NzMyMzhhZTE1NDNiYzY0MmRlNGJhY2Q1NjlmNTk1NGFlN2VlXS98/Q==: --dhchap-ctrl-secret DHHC-1:03:NmM0ODExMWYzNDJlOTA1ZjgyNTliZThhZmQ2NGY1YTMwZGFmYmVkMTY3MmFiNjRkZDdiZTAwNmRhZDYxYmI1N4Bp9x8=: 00:17:45.334 18:03:38 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:45.334 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:45.334 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:45.334 18:03:39 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:45.334 18:03:39 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:45.334 18:03:39 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:45.334 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:45.334 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:17:45.334 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:17:45.593 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 ffdhe8192 1 00:17:45.593 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:45.593 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:45.593 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe8192 00:17:45.593 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key1 00:17:45.593 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:45.593 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:45.593 18:03:39 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:45.593 18:03:39 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:45.593 18:03:39 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:45.593 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:45.593 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:46.162 00:17:46.162 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:46.162 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:46.162 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:46.162 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:46.162 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:46.162 18:03:39 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:46.162 18:03:39 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:46.162 18:03:39 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:46.162 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:46.162 { 00:17:46.162 "cntlid": 139, 00:17:46.162 "qid": 0, 00:17:46.162 "state": "enabled", 00:17:46.162 "thread": "nvmf_tgt_poll_group_000", 00:17:46.162 "listen_address": { 00:17:46.162 "trtype": "TCP", 00:17:46.162 "adrfam": "IPv4", 00:17:46.162 "traddr": "10.0.0.2", 00:17:46.162 "trsvcid": "4420" 00:17:46.162 }, 00:17:46.162 "peer_address": { 00:17:46.162 "trtype": "TCP", 00:17:46.162 "adrfam": "IPv4", 00:17:46.162 "traddr": "10.0.0.1", 00:17:46.162 "trsvcid": "60002" 00:17:46.162 }, 00:17:46.162 "auth": { 00:17:46.162 "state": "completed", 00:17:46.162 "digest": "sha512", 00:17:46.162 "dhgroup": "ffdhe8192" 00:17:46.162 } 00:17:46.162 } 00:17:46.162 ]' 00:17:46.162 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:46.421 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:46.421 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:46.421 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:46.421 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:46.421 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:46.421 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:46.421 18:03:39 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:46.680 18:03:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:01:ZDMzNjZhODE5ZDkwZjc1YzJlMWRhMzJmNzNkZWZjM2TGg2/J: --dhchap-ctrl-secret DHHC-1:02:OWFmZjc0MDk4ZDJmZWMyYWFjNmJmNmM1MzZkMjUwMDZlY2RmOGMxNjQxNzZkZDdloY1p5Q==: 00:17:47.247 18:03:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:47.247 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:47.247 18:03:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:47.247 18:03:40 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:47.247 18:03:40 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:47.247 18:03:40 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:47.247 18:03:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:47.247 18:03:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:17:47.248 18:03:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:17:47.248 18:03:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 ffdhe8192 2 00:17:47.248 18:03:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:47.248 18:03:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:47.248 18:03:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe8192 00:17:47.248 18:03:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key2 00:17:47.248 18:03:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:47.248 18:03:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:47.248 18:03:40 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:47.248 18:03:40 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:47.248 18:03:40 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:47.248 18:03:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:47.248 18:03:40 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:17:47.816 00:17:47.816 18:03:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:47.816 18:03:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:47.816 18:03:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:48.075 18:03:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:48.075 18:03:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:48.075 18:03:41 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:48.075 18:03:41 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:48.075 18:03:41 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:48.075 18:03:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:48.075 { 00:17:48.075 "cntlid": 141, 00:17:48.075 "qid": 0, 00:17:48.075 "state": "enabled", 00:17:48.075 "thread": "nvmf_tgt_poll_group_000", 00:17:48.075 "listen_address": { 00:17:48.075 "trtype": "TCP", 00:17:48.075 "adrfam": "IPv4", 00:17:48.075 "traddr": "10.0.0.2", 00:17:48.075 "trsvcid": "4420" 00:17:48.075 }, 00:17:48.075 "peer_address": { 00:17:48.075 "trtype": "TCP", 00:17:48.075 "adrfam": "IPv4", 00:17:48.075 "traddr": "10.0.0.1", 00:17:48.075 "trsvcid": "60036" 00:17:48.075 }, 00:17:48.075 "auth": { 00:17:48.075 "state": "completed", 00:17:48.075 "digest": "sha512", 00:17:48.075 "dhgroup": "ffdhe8192" 00:17:48.075 } 00:17:48.075 } 00:17:48.075 ]' 00:17:48.075 18:03:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:48.075 18:03:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:48.075 18:03:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:48.075 18:03:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:48.075 18:03:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:48.075 18:03:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:48.075 18:03:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:48.075 18:03:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:48.335 18:03:41 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:02:M2JjZDM2NWNhZmQ5MTRjN2UwY2Q2YjhiOWUwYmE1MzZiNzNiZDYwNjZhZGJkYjM3IIKM4Q==: --dhchap-ctrl-secret DHHC-1:01:OTFlOTkyN2M2NzZiZWZkNWRjOTE4ZmUxNTUyYjI3ODS55lrp: 00:17:48.941 18:03:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:48.941 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:48.941 18:03:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:48.941 18:03:42 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:48.941 18:03:42 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:48.941 18:03:42 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:48.941 18:03:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@93 -- # for keyid in "${!keys[@]}" 00:17:48.941 18:03:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@94 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:17:48.941 18:03:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:17:49.208 18:03:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@96 -- # connect_authenticate sha512 ffdhe8192 3 00:17:49.208 18:03:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:49.208 18:03:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:49.208 18:03:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe8192 00:17:49.208 18:03:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key3 00:17:49.208 18:03:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:49.208 18:03:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key3 00:17:49.208 18:03:42 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:49.208 18:03:42 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:49.208 18:03:42 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:49.208 18:03:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:49.208 18:03:42 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:49.467 00:17:49.467 18:03:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:49.467 18:03:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:49.467 18:03:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:49.726 18:03:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:49.726 18:03:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:49.726 18:03:43 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:49.726 18:03:43 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:49.726 18:03:43 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:49.726 18:03:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:49.726 { 00:17:49.726 "cntlid": 143, 00:17:49.726 "qid": 0, 00:17:49.726 "state": "enabled", 00:17:49.726 "thread": "nvmf_tgt_poll_group_000", 00:17:49.726 "listen_address": { 00:17:49.726 "trtype": "TCP", 00:17:49.726 "adrfam": "IPv4", 00:17:49.726 "traddr": "10.0.0.2", 00:17:49.726 "trsvcid": "4420" 00:17:49.726 }, 00:17:49.726 "peer_address": { 00:17:49.726 "trtype": "TCP", 00:17:49.726 "adrfam": "IPv4", 00:17:49.726 "traddr": "10.0.0.1", 00:17:49.726 "trsvcid": "60060" 00:17:49.726 }, 00:17:49.726 "auth": { 00:17:49.726 "state": "completed", 00:17:49.726 "digest": "sha512", 00:17:49.726 "dhgroup": "ffdhe8192" 00:17:49.726 } 00:17:49.726 } 00:17:49.726 ]' 00:17:49.726 18:03:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:49.726 18:03:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:49.726 18:03:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:49.726 18:03:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:49.726 18:03:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:49.985 18:03:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:49.985 18:03:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:49.985 18:03:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:49.985 18:03:43 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:03:MDkyZTk1N2U3YjVlMDJjNThmOTAwMDMzY2YzZWQ3ZjZiYTk2ZDA3MzczYTM5MTgyNDYxMjY5ZjBjMDlhNGMzZRaivds=: 00:17:50.552 18:03:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:50.552 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:50.553 18:03:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:50.553 18:03:44 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:50.553 18:03:44 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:50.553 18:03:44 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:50.553 18:03:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@102 -- # IFS=, 00:17:50.553 18:03:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@103 -- # printf %s sha256,sha384,sha512 00:17:50.553 18:03:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@102 -- # IFS=, 00:17:50.553 18:03:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@103 -- # printf %s null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:17:50.553 18:03:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@102 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:17:50.553 18:03:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:17:50.811 18:03:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@114 -- # connect_authenticate sha512 ffdhe8192 0 00:17:50.811 18:03:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:50.811 18:03:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:50.811 18:03:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe8192 00:17:50.811 18:03:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key0 00:17:50.811 18:03:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:50.811 18:03:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:50.811 18:03:44 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:50.811 18:03:44 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:50.811 18:03:44 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:50.811 18:03:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:50.811 18:03:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:17:51.387 00:17:51.387 18:03:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:51.387 18:03:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:51.387 18:03:44 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:51.387 18:03:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:51.387 18:03:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:51.387 18:03:45 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:51.387 18:03:45 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:51.646 18:03:45 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:51.646 18:03:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:51.646 { 00:17:51.646 "cntlid": 145, 00:17:51.646 "qid": 0, 00:17:51.646 "state": "enabled", 00:17:51.646 "thread": "nvmf_tgt_poll_group_000", 00:17:51.646 "listen_address": { 00:17:51.646 "trtype": "TCP", 00:17:51.646 "adrfam": "IPv4", 00:17:51.646 "traddr": "10.0.0.2", 00:17:51.646 "trsvcid": "4420" 00:17:51.646 }, 00:17:51.646 "peer_address": { 00:17:51.646 "trtype": "TCP", 00:17:51.646 "adrfam": "IPv4", 00:17:51.646 "traddr": "10.0.0.1", 00:17:51.646 "trsvcid": "60084" 00:17:51.646 }, 00:17:51.646 "auth": { 00:17:51.646 "state": "completed", 00:17:51.646 "digest": "sha512", 00:17:51.646 "dhgroup": "ffdhe8192" 00:17:51.646 } 00:17:51.646 } 00:17:51.646 ]' 00:17:51.646 18:03:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:51.646 18:03:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:51.646 18:03:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:51.646 18:03:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:51.646 18:03:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:51.646 18:03:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:51.646 18:03:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:51.646 18:03:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:51.905 18:03:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:00:ZDc0NTQyMWE2OWE3NzMyMzhhZTE1NDNiYzY0MmRlNGJhY2Q1NjlmNTk1NGFlN2VlXS98/Q==: --dhchap-ctrl-secret DHHC-1:03:NmM0ODExMWYzNDJlOTA1ZjgyNTliZThhZmQ2NGY1YTMwZGFmYmVkMTY3MmFiNjRkZDdiZTAwNmRhZDYxYmI1N4Bp9x8=: 00:17:52.473 18:03:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:52.473 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:52.473 18:03:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:52.473 18:03:45 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:52.473 18:03:45 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:52.473 18:03:45 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:52.473 18:03:45 nvmf_tcp.nvmf_auth_target -- target/auth.sh@117 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 00:17:52.473 18:03:45 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:52.473 18:03:45 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:52.473 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:52.473 18:03:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@118 -- # NOT hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 00:17:52.473 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@648 -- # local es=0 00:17:52.473 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@650 -- # valid_exec_arg hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 00:17:52.473 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@636 -- # local arg=hostrpc 00:17:52.473 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:17:52.473 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@640 -- # type -t hostrpc 00:17:52.473 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:17:52.473 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@651 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 00:17:52.473 18:03:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key2 00:17:52.731 request: 00:17:52.731 { 00:17:52.731 "name": "nvme0", 00:17:52.731 "trtype": "tcp", 00:17:52.731 "traddr": "10.0.0.2", 00:17:52.731 "adrfam": "ipv4", 00:17:52.731 "trsvcid": "4420", 00:17:52.731 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:17:52.731 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562", 00:17:52.731 "prchk_reftag": false, 00:17:52.731 "prchk_guard": false, 00:17:52.731 "hdgst": false, 00:17:52.731 "ddgst": false, 00:17:52.731 "dhchap_key": "key2", 00:17:52.731 "method": "bdev_nvme_attach_controller", 00:17:52.731 "req_id": 1 00:17:52.731 } 00:17:52.731 Got JSON-RPC error response 00:17:52.731 response: 00:17:52.731 { 00:17:52.731 "code": -5, 00:17:52.731 "message": "Input/output error" 00:17:52.731 } 00:17:52.731 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@651 -- # es=1 00:17:52.731 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:17:52.731 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:17:52.731 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:17:52.731 18:03:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@121 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:52.731 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:52.989 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:52.989 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:52.989 18:03:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@124 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:52.989 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:52.989 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:52.989 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:52.989 18:03:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@125 -- # NOT hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:17:52.989 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@648 -- # local es=0 00:17:52.989 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@650 -- # valid_exec_arg hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:17:52.989 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@636 -- # local arg=hostrpc 00:17:52.989 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:17:52.989 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@640 -- # type -t hostrpc 00:17:52.989 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:17:52.989 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@651 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:17:52.990 18:03:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:17:53.248 request: 00:17:53.248 { 00:17:53.248 "name": "nvme0", 00:17:53.248 "trtype": "tcp", 00:17:53.248 "traddr": "10.0.0.2", 00:17:53.248 "adrfam": "ipv4", 00:17:53.248 "trsvcid": "4420", 00:17:53.248 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:17:53.248 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562", 00:17:53.248 "prchk_reftag": false, 00:17:53.248 "prchk_guard": false, 00:17:53.248 "hdgst": false, 00:17:53.248 "ddgst": false, 00:17:53.248 "dhchap_key": "key1", 00:17:53.248 "dhchap_ctrlr_key": "ckey2", 00:17:53.248 "method": "bdev_nvme_attach_controller", 00:17:53.248 "req_id": 1 00:17:53.248 } 00:17:53.248 Got JSON-RPC error response 00:17:53.248 response: 00:17:53.248 { 00:17:53.248 "code": -5, 00:17:53.248 "message": "Input/output error" 00:17:53.248 } 00:17:53.248 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@651 -- # es=1 00:17:53.248 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:17:53.248 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:17:53.248 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:17:53.249 18:03:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@128 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:53.249 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:53.249 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:53.249 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:53.249 18:03:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@131 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key1 00:17:53.249 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:53.249 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:53.249 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:53.249 18:03:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@132 -- # NOT hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:53.249 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@648 -- # local es=0 00:17:53.249 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@650 -- # valid_exec_arg hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:53.249 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@636 -- # local arg=hostrpc 00:17:53.249 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:17:53.249 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@640 -- # type -t hostrpc 00:17:53.249 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:17:53.249 18:03:46 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@651 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:53.249 18:03:46 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:17:53.817 request: 00:17:53.817 { 00:17:53.817 "name": "nvme0", 00:17:53.817 "trtype": "tcp", 00:17:53.817 "traddr": "10.0.0.2", 00:17:53.817 "adrfam": "ipv4", 00:17:53.817 "trsvcid": "4420", 00:17:53.817 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:17:53.817 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562", 00:17:53.817 "prchk_reftag": false, 00:17:53.817 "prchk_guard": false, 00:17:53.817 "hdgst": false, 00:17:53.817 "ddgst": false, 00:17:53.817 "dhchap_key": "key1", 00:17:53.817 "dhchap_ctrlr_key": "ckey1", 00:17:53.817 "method": "bdev_nvme_attach_controller", 00:17:53.817 "req_id": 1 00:17:53.817 } 00:17:53.817 Got JSON-RPC error response 00:17:53.817 response: 00:17:53.817 { 00:17:53.817 "code": -5, 00:17:53.817 "message": "Input/output error" 00:17:53.817 } 00:17:53.817 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@651 -- # es=1 00:17:53.817 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:17:53.817 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:17:53.817 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:17:53.817 18:03:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@135 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:53.817 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:53.817 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:53.817 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:53.817 18:03:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@138 -- # killprocess 583809 00:17:53.817 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@948 -- # '[' -z 583809 ']' 00:17:53.817 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@952 -- # kill -0 583809 00:17:53.817 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@953 -- # uname 00:17:53.817 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:17:53.817 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 583809 00:17:53.817 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:17:53.817 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:17:53.817 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@966 -- # echo 'killing process with pid 583809' 00:17:53.817 killing process with pid 583809 00:17:53.817 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@967 -- # kill 583809 00:17:53.817 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@972 -- # wait 583809 00:17:54.076 18:03:47 nvmf_tcp.nvmf_auth_target -- target/auth.sh@139 -- # nvmfappstart --wait-for-rpc -L nvmf_auth 00:17:54.076 18:03:47 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:17:54.076 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@722 -- # xtrace_disable 00:17:54.076 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:54.076 18:03:47 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@481 -- # nvmfpid=604764 00:17:54.076 18:03:47 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc -L nvmf_auth 00:17:54.076 18:03:47 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@482 -- # waitforlisten 604764 00:17:54.076 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@829 -- # '[' -z 604764 ']' 00:17:54.076 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:54.076 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@834 -- # local max_retries=100 00:17:54.076 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:54.076 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@838 -- # xtrace_disable 00:17:54.076 18:03:47 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:55.014 18:03:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:17:55.014 18:03:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@862 -- # return 0 00:17:55.014 18:03:48 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:17:55.014 18:03:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@728 -- # xtrace_disable 00:17:55.014 18:03:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:55.014 18:03:48 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:17:55.014 18:03:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@140 -- # trap 'dumplogs; cleanup' SIGINT SIGTERM EXIT 00:17:55.014 18:03:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@142 -- # waitforlisten 604764 00:17:55.014 18:03:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@829 -- # '[' -z 604764 ']' 00:17:55.014 18:03:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:17:55.014 18:03:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@834 -- # local max_retries=100 00:17:55.014 18:03:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:17:55.014 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:17:55.014 18:03:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@838 -- # xtrace_disable 00:17:55.014 18:03:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:55.014 18:03:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:17:55.014 18:03:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@862 -- # return 0 00:17:55.014 18:03:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@143 -- # rpc_cmd 00:17:55.014 18:03:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:55.014 18:03:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:55.274 18:03:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:55.274 18:03:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@153 -- # connect_authenticate sha512 ffdhe8192 3 00:17:55.274 18:03:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@34 -- # local digest dhgroup key ckey qpairs 00:17:55.274 18:03:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # digest=sha512 00:17:55.274 18:03:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # dhgroup=ffdhe8192 00:17:55.274 18:03:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@36 -- # key=key3 00:17:55.274 18:03:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@37 -- # ckey=(${ckeys[$3]:+--dhchap-ctrlr-key "ckey$3"}) 00:17:55.274 18:03:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@39 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key3 00:17:55.274 18:03:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:55.274 18:03:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:55.274 18:03:48 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:55.274 18:03:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@40 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:55.274 18:03:48 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:55.842 00:17:55.842 18:03:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # hostrpc bdev_nvme_get_controllers 00:17:55.842 18:03:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # jq -r '.[].name' 00:17:55.842 18:03:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:55.842 18:03:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@44 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:55.842 18:03:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # rpc_cmd nvmf_subsystem_get_qpairs nqn.2024-03.io.spdk:cnode0 00:17:55.842 18:03:49 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:55.842 18:03:49 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:55.842 18:03:49 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:55.842 18:03:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@45 -- # qpairs='[ 00:17:55.842 { 00:17:55.842 "cntlid": 1, 00:17:55.842 "qid": 0, 00:17:55.842 "state": "enabled", 00:17:55.842 "thread": "nvmf_tgt_poll_group_000", 00:17:55.842 "listen_address": { 00:17:55.842 "trtype": "TCP", 00:17:55.842 "adrfam": "IPv4", 00:17:55.842 "traddr": "10.0.0.2", 00:17:55.842 "trsvcid": "4420" 00:17:55.842 }, 00:17:55.842 "peer_address": { 00:17:55.842 "trtype": "TCP", 00:17:55.842 "adrfam": "IPv4", 00:17:55.842 "traddr": "10.0.0.1", 00:17:55.842 "trsvcid": "33370" 00:17:55.842 }, 00:17:55.842 "auth": { 00:17:55.842 "state": "completed", 00:17:55.842 "digest": "sha512", 00:17:55.842 "dhgroup": "ffdhe8192" 00:17:55.842 } 00:17:55.842 } 00:17:55.842 ]' 00:17:55.842 18:03:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # jq -r '.[0].auth.digest' 00:17:55.842 18:03:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@46 -- # [[ sha512 == \s\h\a\5\1\2 ]] 00:17:55.842 18:03:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # jq -r '.[0].auth.dhgroup' 00:17:56.100 18:03:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@47 -- # [[ ffdhe8192 == \f\f\d\h\e\8\1\9\2 ]] 00:17:56.100 18:03:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # jq -r '.[0].auth.state' 00:17:56.100 18:03:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@48 -- # [[ completed == \c\o\m\p\l\e\t\e\d ]] 00:17:56.101 18:03:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@49 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:56.101 18:03:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:56.101 18:03:49 nvmf_tcp.nvmf_auth_target -- target/auth.sh@52 -- # nvme connect -t tcp -a 10.0.0.2 -n nqn.2024-03.io.spdk:cnode0 -i 1 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid 80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-secret DHHC-1:03:MDkyZTk1N2U3YjVlMDJjNThmOTAwMDMzY2YzZWQ3ZjZiYTk2ZDA3MzczYTM5MTgyNDYxMjY5ZjBjMDlhNGMzZRaivds=: 00:17:56.669 18:03:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@55 -- # nvme disconnect -n nqn.2024-03.io.spdk:cnode0 00:17:56.669 NQN:nqn.2024-03.io.spdk:cnode0 disconnected 1 controller(s) 00:17:56.669 18:03:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@56 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:56.669 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:56.669 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:56.669 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:56.669 18:03:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@156 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --dhchap-key key3 00:17:56.669 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:56.669 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:56.669 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:56.669 18:03:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@157 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256 00:17:56.669 18:03:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256 00:17:56.928 18:03:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@158 -- # NOT hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:56.928 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@648 -- # local es=0 00:17:56.928 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@650 -- # valid_exec_arg hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:56.928 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@636 -- # local arg=hostrpc 00:17:56.928 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:17:56.928 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@640 -- # type -t hostrpc 00:17:56.928 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:17:56.928 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@651 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:56.928 18:03:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:57.187 request: 00:17:57.187 { 00:17:57.187 "name": "nvme0", 00:17:57.187 "trtype": "tcp", 00:17:57.187 "traddr": "10.0.0.2", 00:17:57.187 "adrfam": "ipv4", 00:17:57.187 "trsvcid": "4420", 00:17:57.187 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:17:57.187 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562", 00:17:57.187 "prchk_reftag": false, 00:17:57.187 "prchk_guard": false, 00:17:57.187 "hdgst": false, 00:17:57.187 "ddgst": false, 00:17:57.187 "dhchap_key": "key3", 00:17:57.187 "method": "bdev_nvme_attach_controller", 00:17:57.187 "req_id": 1 00:17:57.187 } 00:17:57.187 Got JSON-RPC error response 00:17:57.187 response: 00:17:57.187 { 00:17:57.187 "code": -5, 00:17:57.187 "message": "Input/output error" 00:17:57.187 } 00:17:57.187 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@651 -- # es=1 00:17:57.187 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:17:57.187 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:17:57.187 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:17:57.187 18:03:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@163 -- # IFS=, 00:17:57.187 18:03:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@164 -- # printf %s sha256,sha384,sha512 00:17:57.187 18:03:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@163 -- # hostrpc bdev_nvme_set_options --dhchap-dhgroups ffdhe2048 --dhchap-digests sha256,sha384,sha512 00:17:57.187 18:03:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-dhgroups ffdhe2048 --dhchap-digests sha256,sha384,sha512 00:17:57.187 18:03:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@169 -- # NOT hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:57.187 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@648 -- # local es=0 00:17:57.187 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@650 -- # valid_exec_arg hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:57.187 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@636 -- # local arg=hostrpc 00:17:57.187 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:17:57.187 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@640 -- # type -t hostrpc 00:17:57.187 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:17:57.187 18:03:50 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@651 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:57.187 18:03:50 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key3 00:17:57.446 request: 00:17:57.446 { 00:17:57.446 "name": "nvme0", 00:17:57.446 "trtype": "tcp", 00:17:57.446 "traddr": "10.0.0.2", 00:17:57.446 "adrfam": "ipv4", 00:17:57.446 "trsvcid": "4420", 00:17:57.446 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:17:57.446 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562", 00:17:57.446 "prchk_reftag": false, 00:17:57.446 "prchk_guard": false, 00:17:57.446 "hdgst": false, 00:17:57.446 "ddgst": false, 00:17:57.446 "dhchap_key": "key3", 00:17:57.446 "method": "bdev_nvme_attach_controller", 00:17:57.446 "req_id": 1 00:17:57.446 } 00:17:57.446 Got JSON-RPC error response 00:17:57.446 response: 00:17:57.446 { 00:17:57.446 "code": -5, 00:17:57.446 "message": "Input/output error" 00:17:57.446 } 00:17:57.446 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@651 -- # es=1 00:17:57.446 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:17:57.446 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:17:57.446 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:17:57.446 18:03:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@175 -- # IFS=, 00:17:57.446 18:03:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@176 -- # printf %s sha256,sha384,sha512 00:17:57.446 18:03:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@175 -- # IFS=, 00:17:57.446 18:03:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@176 -- # printf %s null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:17:57.446 18:03:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@175 -- # hostrpc bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:17:57.446 18:03:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups null,ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:17:57.704 18:03:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@186 -- # rpc_cmd nvmf_subsystem_remove_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:57.704 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:57.704 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:57.704 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:57.704 18:03:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@187 -- # rpc_cmd nvmf_subsystem_add_host nqn.2024-03.io.spdk:cnode0 nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:17:57.704 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@559 -- # xtrace_disable 00:17:57.704 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:17:57.704 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:17:57.704 18:03:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@188 -- # NOT hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:17:57.704 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@648 -- # local es=0 00:17:57.704 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@650 -- # valid_exec_arg hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:17:57.704 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@636 -- # local arg=hostrpc 00:17:57.704 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:17:57.704 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@640 -- # type -t hostrpc 00:17:57.705 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:17:57.705 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@651 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:17:57.705 18:03:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key key1 00:17:57.963 request: 00:17:57.963 { 00:17:57.963 "name": "nvme0", 00:17:57.963 "trtype": "tcp", 00:17:57.963 "traddr": "10.0.0.2", 00:17:57.963 "adrfam": "ipv4", 00:17:57.963 "trsvcid": "4420", 00:17:57.963 "subnqn": "nqn.2024-03.io.spdk:cnode0", 00:17:57.963 "hostnqn": "nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562", 00:17:57.963 "prchk_reftag": false, 00:17:57.963 "prchk_guard": false, 00:17:57.963 "hdgst": false, 00:17:57.963 "ddgst": false, 00:17:57.963 "dhchap_key": "key0", 00:17:57.963 "dhchap_ctrlr_key": "key1", 00:17:57.963 "method": "bdev_nvme_attach_controller", 00:17:57.963 "req_id": 1 00:17:57.963 } 00:17:57.963 Got JSON-RPC error response 00:17:57.963 response: 00:17:57.963 { 00:17:57.963 "code": -5, 00:17:57.963 "message": "Input/output error" 00:17:57.963 } 00:17:57.963 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@651 -- # es=1 00:17:57.963 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:17:57.963 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:17:57.963 18:03:51 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:17:57.963 18:03:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@192 -- # hostrpc bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 00:17:57.963 18:03:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.2 -s 4420 -q nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 -n nqn.2024-03.io.spdk:cnode0 --dhchap-key key0 00:17:57.963 00:17:58.221 18:03:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@195 -- # hostrpc bdev_nvme_get_controllers 00:17:58.221 18:03:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@195 -- # jq -r '.[].name' 00:17:58.221 18:03:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_get_controllers 00:17:58.221 18:03:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@195 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:17:58.221 18:03:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@196 -- # hostrpc bdev_nvme_detach_controller nvme0 00:17:58.221 18:03:51 nvmf_tcp.nvmf_auth_target -- target/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/host.sock bdev_nvme_detach_controller nvme0 00:17:58.480 18:03:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@198 -- # trap - SIGINT SIGTERM EXIT 00:17:58.480 18:03:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@199 -- # cleanup 00:17:58.480 18:03:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@21 -- # killprocess 583868 00:17:58.480 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@948 -- # '[' -z 583868 ']' 00:17:58.480 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@952 -- # kill -0 583868 00:17:58.480 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@953 -- # uname 00:17:58.480 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:17:58.480 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 583868 00:17:58.480 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:17:58.480 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:17:58.481 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@966 -- # echo 'killing process with pid 583868' 00:17:58.481 killing process with pid 583868 00:17:58.481 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@967 -- # kill 583868 00:17:58.481 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@972 -- # wait 583868 00:17:58.739 18:03:52 nvmf_tcp.nvmf_auth_target -- target/auth.sh@22 -- # nvmftestfini 00:17:58.739 18:03:52 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@488 -- # nvmfcleanup 00:17:58.739 18:03:52 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@117 -- # sync 00:17:58.739 18:03:52 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:17:58.739 18:03:52 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@120 -- # set +e 00:17:58.739 18:03:52 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@121 -- # for i in {1..20} 00:17:58.739 18:03:52 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:17:58.739 rmmod nvme_tcp 00:17:58.739 rmmod nvme_fabrics 00:17:58.739 rmmod nvme_keyring 00:17:58.739 18:03:52 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:17:58.739 18:03:52 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@124 -- # set -e 00:17:58.739 18:03:52 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@125 -- # return 0 00:17:58.739 18:03:52 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@489 -- # '[' -n 604764 ']' 00:17:58.739 18:03:52 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@490 -- # killprocess 604764 00:17:58.739 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@948 -- # '[' -z 604764 ']' 00:17:58.739 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@952 -- # kill -0 604764 00:17:58.998 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@953 -- # uname 00:17:58.998 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:17:58.998 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 604764 00:17:58.998 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:17:58.998 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:17:58.998 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@966 -- # echo 'killing process with pid 604764' 00:17:58.998 killing process with pid 604764 00:17:58.998 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@967 -- # kill 604764 00:17:58.998 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@972 -- # wait 604764 00:17:58.998 18:03:52 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:17:58.999 18:03:52 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:17:58.999 18:03:52 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:17:58.999 18:03:52 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:17:58.999 18:03:52 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@278 -- # remove_spdk_ns 00:17:58.999 18:03:52 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:17:58.999 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:17:58.999 18:03:52 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:18:01.535 18:03:54 nvmf_tcp.nvmf_auth_target -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:18:01.535 18:03:54 nvmf_tcp.nvmf_auth_target -- target/auth.sh@23 -- # rm -f /tmp/spdk.key-null.mBL /tmp/spdk.key-sha256.aI0 /tmp/spdk.key-sha384.pOl /tmp/spdk.key-sha512.IyM /tmp/spdk.key-sha512.AEk /tmp/spdk.key-sha384.TJQ /tmp/spdk.key-sha256.7IT '' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvme-auth.log /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf-auth.log 00:18:01.535 00:18:01.535 real 2m10.823s 00:18:01.535 user 5m0.308s 00:18:01.535 sys 0m20.300s 00:18:01.535 18:03:54 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@1124 -- # xtrace_disable 00:18:01.535 18:03:54 nvmf_tcp.nvmf_auth_target -- common/autotest_common.sh@10 -- # set +x 00:18:01.535 ************************************ 00:18:01.535 END TEST nvmf_auth_target 00:18:01.535 ************************************ 00:18:01.535 18:03:54 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:18:01.535 18:03:54 nvmf_tcp -- nvmf/nvmf.sh@59 -- # '[' tcp = tcp ']' 00:18:01.535 18:03:54 nvmf_tcp -- nvmf/nvmf.sh@60 -- # run_test nvmf_bdevio_no_huge /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp --no-hugepages 00:18:01.535 18:03:54 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 4 -le 1 ']' 00:18:01.535 18:03:54 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:18:01.535 18:03:54 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:18:01.535 ************************************ 00:18:01.535 START TEST nvmf_bdevio_no_huge 00:18:01.535 ************************************ 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevio.sh --transport=tcp --no-hugepages 00:18:01.535 * Looking for test storage... 00:18:01.535 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- target/bdevio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@7 -- # uname -s 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- paths/export.sh@5 -- # export PATH 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@47 -- # : 0 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@51 -- # have_pci_nics=0 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- target/bdevio.sh@11 -- # MALLOC_BDEV_SIZE=64 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- target/bdevio.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- target/bdevio.sh@14 -- # nvmftestinit 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@448 -- # prepare_net_devs 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@410 -- # local -g is_hw=no 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@412 -- # remove_spdk_ns 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@285 -- # xtrace_disable 00:18:01.535 18:03:54 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@291 -- # pci_devs=() 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@291 -- # local -a pci_devs 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@292 -- # pci_net_devs=() 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@293 -- # pci_drivers=() 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@293 -- # local -A pci_drivers 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@295 -- # net_devs=() 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@295 -- # local -ga net_devs 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@296 -- # e810=() 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@296 -- # local -ga e810 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@297 -- # x722=() 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@297 -- # local -ga x722 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@298 -- # mlx=() 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@298 -- # local -ga mlx 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:18:06.839 Found 0000:86:00.0 (0x8086 - 0x159b) 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:18:06.839 Found 0000:86:00.1 (0x8086 - 0x159b) 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@390 -- # [[ up == up ]] 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:18:06.839 Found net devices under 0000:86:00.0: cvl_0_0 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@390 -- # [[ up == up ]] 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:18:06.839 Found net devices under 0000:86:00.1: cvl_0_1 00:18:06.839 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@414 -- # is_hw=yes 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:18:06.840 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:18:06.840 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.172 ms 00:18:06.840 00:18:06.840 --- 10.0.0.2 ping statistics --- 00:18:06.840 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:18:06.840 rtt min/avg/max/mdev = 0.172/0.172/0.172/0.000 ms 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:18:06.840 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:18:06.840 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.239 ms 00:18:06.840 00:18:06.840 --- 10.0.0.1 ping statistics --- 00:18:06.840 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:18:06.840 rtt min/avg/max/mdev = 0.239/0.239/0.239/0.000 ms 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@422 -- # return 0 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:18:06.840 18:03:59 nvmf_tcp.nvmf_bdevio_no_huge -- target/bdevio.sh@16 -- # nvmfappstart -m 0x78 00:18:06.840 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:18:06.840 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@722 -- # xtrace_disable 00:18:06.840 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:06.840 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@481 -- # nvmfpid=609023 00:18:06.840 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@482 -- # waitforlisten 609023 00:18:06.840 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --no-huge -s 1024 -m 0x78 00:18:06.840 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@829 -- # '[' -z 609023 ']' 00:18:06.840 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:06.840 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@834 -- # local max_retries=100 00:18:06.840 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:06.840 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:06.840 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@838 -- # xtrace_disable 00:18:06.840 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:06.840 [2024-07-15 18:04:00.053845] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:18:06.840 [2024-07-15 18:04:00.053890] [ DPDK EAL parameters: nvmf -c 0x78 -m 1024 --no-huge --iova-mode=va --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --file-prefix=spdk0 --proc-type=auto ] 00:18:06.840 [2024-07-15 18:04:00.117502] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:18:06.840 [2024-07-15 18:04:00.200542] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:18:06.840 [2024-07-15 18:04:00.200580] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:18:06.840 [2024-07-15 18:04:00.200587] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:18:06.840 [2024-07-15 18:04:00.200593] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:18:06.840 [2024-07-15 18:04:00.200598] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:18:06.840 [2024-07-15 18:04:00.200713] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 4 00:18:06.840 [2024-07-15 18:04:00.200817] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 5 00:18:06.840 [2024-07-15 18:04:00.200903] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:18:06.840 [2024-07-15 18:04:00.200904] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 6 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@862 -- # return 0 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@728 -- # xtrace_disable 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- target/bdevio.sh@18 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@559 -- # xtrace_disable 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:07.407 [2024-07-15 18:04:00.908178] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- target/bdevio.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@559 -- # xtrace_disable 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:07.407 Malloc0 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- target/bdevio.sh@20 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@559 -- # xtrace_disable 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- target/bdevio.sh@21 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@559 -- # xtrace_disable 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- target/bdevio.sh@22 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@559 -- # xtrace_disable 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:07.407 [2024-07-15 18:04:00.952477] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- target/bdevio.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/bdev/bdevio/bdevio --json /dev/fd/62 --no-huge -s 1024 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- target/bdevio.sh@24 -- # gen_nvmf_target_json 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@532 -- # config=() 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@532 -- # local subsystem config 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:18:07.407 { 00:18:07.407 "params": { 00:18:07.407 "name": "Nvme$subsystem", 00:18:07.407 "trtype": "$TEST_TRANSPORT", 00:18:07.407 "traddr": "$NVMF_FIRST_TARGET_IP", 00:18:07.407 "adrfam": "ipv4", 00:18:07.407 "trsvcid": "$NVMF_PORT", 00:18:07.407 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:18:07.407 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:18:07.407 "hdgst": ${hdgst:-false}, 00:18:07.407 "ddgst": ${ddgst:-false} 00:18:07.407 }, 00:18:07.407 "method": "bdev_nvme_attach_controller" 00:18:07.407 } 00:18:07.407 EOF 00:18:07.407 )") 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@554 -- # cat 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@556 -- # jq . 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@557 -- # IFS=, 00:18:07.407 18:04:00 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:18:07.407 "params": { 00:18:07.407 "name": "Nvme1", 00:18:07.407 "trtype": "tcp", 00:18:07.407 "traddr": "10.0.0.2", 00:18:07.407 "adrfam": "ipv4", 00:18:07.407 "trsvcid": "4420", 00:18:07.407 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:18:07.407 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:18:07.407 "hdgst": false, 00:18:07.407 "ddgst": false 00:18:07.407 }, 00:18:07.407 "method": "bdev_nvme_attach_controller" 00:18:07.407 }' 00:18:07.407 [2024-07-15 18:04:01.001595] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:18:07.407 [2024-07-15 18:04:01.001643] [ DPDK EAL parameters: bdevio --no-shconf -c 0x7 -m 1024 --no-huge --iova-mode=va --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --file-prefix=spdk_pid609074 ] 00:18:07.407 [2024-07-15 18:04:01.059331] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:18:07.665 [2024-07-15 18:04:01.145521] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:18:07.665 [2024-07-15 18:04:01.145617] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:18:07.665 [2024-07-15 18:04:01.145617] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:18:07.665 I/O targets: 00:18:07.665 Nvme1n1: 131072 blocks of 512 bytes (64 MiB) 00:18:07.665 00:18:07.665 00:18:07.665 CUnit - A unit testing framework for C - Version 2.1-3 00:18:07.665 http://cunit.sourceforge.net/ 00:18:07.665 00:18:07.665 00:18:07.665 Suite: bdevio tests on: Nvme1n1 00:18:07.665 Test: blockdev write read block ...passed 00:18:07.922 Test: blockdev write zeroes read block ...passed 00:18:07.922 Test: blockdev write zeroes read no split ...passed 00:18:07.922 Test: blockdev write zeroes read split ...passed 00:18:07.922 Test: blockdev write zeroes read split partial ...passed 00:18:07.922 Test: blockdev reset ...[2024-07-15 18:04:01.539647] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:18:07.922 [2024-07-15 18:04:01.539704] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1be8300 (9): Bad file descriptor 00:18:07.922 [2024-07-15 18:04:01.550509] bdev_nvme.c:2067:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:18:07.922 passed 00:18:07.922 Test: blockdev write read 8 blocks ...passed 00:18:07.922 Test: blockdev write read size > 128k ...passed 00:18:07.922 Test: blockdev write read invalid size ...passed 00:18:07.922 Test: blockdev write read offset + nbytes == size of blockdev ...passed 00:18:07.922 Test: blockdev write read offset + nbytes > size of blockdev ...passed 00:18:07.922 Test: blockdev write read max offset ...passed 00:18:08.181 Test: blockdev write read 2 blocks on overlapped address offset ...passed 00:18:08.181 Test: blockdev writev readv 8 blocks ...passed 00:18:08.181 Test: blockdev writev readv 30 x 1block ...passed 00:18:08.181 Test: blockdev writev readv block ...passed 00:18:08.181 Test: blockdev writev readv size > 128k ...passed 00:18:08.181 Test: blockdev writev readv size > 128k in two iovs ...passed 00:18:08.181 Test: blockdev comparev and writev ...[2024-07-15 18:04:01.723394] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:08.181 [2024-07-15 18:04:01.723422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:18:08.181 [2024-07-15 18:04:01.723436] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:08.181 [2024-07-15 18:04:01.723443] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:18:08.181 [2024-07-15 18:04:01.723708] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:08.181 [2024-07-15 18:04:01.723719] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:18:08.181 [2024-07-15 18:04:01.723731] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:08.181 [2024-07-15 18:04:01.723738] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:18:08.181 [2024-07-15 18:04:01.724005] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:08.181 [2024-07-15 18:04:01.724017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:0 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:18:08.181 [2024-07-15 18:04:01.724029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:08.181 [2024-07-15 18:04:01.724038] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:1 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:18:08.181 [2024-07-15 18:04:01.724293] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: COMPARE sqid:1 cid:1 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:08.181 [2024-07-15 18:04:01.724306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMPARE FAILURE (02/85) qid:1 cid:1 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:18:08.181 [2024-07-15 18:04:01.724317] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:0 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x200 00:18:08.181 [2024-07-15 18:04:01.724326] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - FAILED FUSED (00/09) qid:1 cid:0 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:18:08.181 passed 00:18:08.181 Test: blockdev nvme passthru rw ...passed 00:18:08.181 Test: blockdev nvme passthru vendor specific ...[2024-07-15 18:04:01.807651] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:18:08.181 [2024-07-15 18:04:01.807673] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:18:08.181 [2024-07-15 18:04:01.807809] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:18:08.181 [2024-07-15 18:04:01.807820] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:18:08.181 [2024-07-15 18:04:01.807955] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:18:08.181 [2024-07-15 18:04:01.807965] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:18:08.181 [2024-07-15 18:04:01.808101] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:1 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:18:08.181 [2024-07-15 18:04:01.808111] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: INVALID OPCODE (00/01) qid:1 cid:0 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:18:08.181 passed 00:18:08.181 Test: blockdev nvme admin passthru ...passed 00:18:08.181 Test: blockdev copy ...passed 00:18:08.181 00:18:08.181 Run Summary: Type Total Ran Passed Failed Inactive 00:18:08.181 suites 1 1 n/a 0 0 00:18:08.181 tests 23 23 23 0 0 00:18:08.181 asserts 152 152 152 0 n/a 00:18:08.181 00:18:08.181 Elapsed time = 1.063 seconds 00:18:08.439 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- target/bdevio.sh@26 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:18:08.439 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@559 -- # xtrace_disable 00:18:08.439 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:08.439 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:18:08.439 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- target/bdevio.sh@28 -- # trap - SIGINT SIGTERM EXIT 00:18:08.439 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- target/bdevio.sh@30 -- # nvmftestfini 00:18:08.439 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@488 -- # nvmfcleanup 00:18:08.439 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@117 -- # sync 00:18:08.439 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:18:08.439 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@120 -- # set +e 00:18:08.439 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@121 -- # for i in {1..20} 00:18:08.439 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:18:08.439 rmmod nvme_tcp 00:18:08.439 rmmod nvme_fabrics 00:18:08.698 rmmod nvme_keyring 00:18:08.698 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:18:08.698 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@124 -- # set -e 00:18:08.698 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@125 -- # return 0 00:18:08.698 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@489 -- # '[' -n 609023 ']' 00:18:08.698 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@490 -- # killprocess 609023 00:18:08.698 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@948 -- # '[' -z 609023 ']' 00:18:08.698 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@952 -- # kill -0 609023 00:18:08.698 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@953 -- # uname 00:18:08.698 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:18:08.698 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 609023 00:18:08.698 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@954 -- # process_name=reactor_3 00:18:08.698 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@958 -- # '[' reactor_3 = sudo ']' 00:18:08.698 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@966 -- # echo 'killing process with pid 609023' 00:18:08.698 killing process with pid 609023 00:18:08.698 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@967 -- # kill 609023 00:18:08.698 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@972 -- # wait 609023 00:18:08.957 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:18:08.957 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:18:08.957 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:18:08.957 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:18:08.957 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@278 -- # remove_spdk_ns 00:18:08.957 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:18:08.957 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:18:08.957 18:04:02 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:18:11.493 18:04:04 nvmf_tcp.nvmf_bdevio_no_huge -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:18:11.493 00:18:11.493 real 0m9.772s 00:18:11.493 user 0m12.555s 00:18:11.493 sys 0m4.684s 00:18:11.493 18:04:04 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@1124 -- # xtrace_disable 00:18:11.493 18:04:04 nvmf_tcp.nvmf_bdevio_no_huge -- common/autotest_common.sh@10 -- # set +x 00:18:11.493 ************************************ 00:18:11.493 END TEST nvmf_bdevio_no_huge 00:18:11.493 ************************************ 00:18:11.493 18:04:04 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:18:11.493 18:04:04 nvmf_tcp -- nvmf/nvmf.sh@61 -- # run_test nvmf_tls /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/tls.sh --transport=tcp 00:18:11.493 18:04:04 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:18:11.493 18:04:04 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:18:11.493 18:04:04 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:18:11.493 ************************************ 00:18:11.493 START TEST nvmf_tls 00:18:11.493 ************************************ 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/tls.sh --transport=tcp 00:18:11.493 * Looking for test storage... 00:18:11.493 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- target/tls.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@7 -- # uname -s 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- paths/export.sh@5 -- # export PATH 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@47 -- # : 0 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@51 -- # have_pci_nics=0 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- target/tls.sh@12 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- target/tls.sh@62 -- # nvmftestinit 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@448 -- # prepare_net_devs 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@410 -- # local -g is_hw=no 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@412 -- # remove_spdk_ns 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- nvmf/common.sh@285 -- # xtrace_disable 00:18:11.493 18:04:04 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@291 -- # pci_devs=() 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@291 -- # local -a pci_devs 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@292 -- # pci_net_devs=() 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@293 -- # pci_drivers=() 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@293 -- # local -A pci_drivers 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@295 -- # net_devs=() 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@295 -- # local -ga net_devs 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@296 -- # e810=() 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@296 -- # local -ga e810 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@297 -- # x722=() 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@297 -- # local -ga x722 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@298 -- # mlx=() 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@298 -- # local -ga mlx 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:18:16.770 Found 0000:86:00.0 (0x8086 - 0x159b) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:18:16.770 Found 0000:86:00.1 (0x8086 - 0x159b) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@390 -- # [[ up == up ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:18:16.770 Found net devices under 0000:86:00.0: cvl_0_0 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@390 -- # [[ up == up ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:18:16.770 Found net devices under 0000:86:00.1: cvl_0_1 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@414 -- # is_hw=yes 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:18:16.770 18:04:09 nvmf_tcp.nvmf_tls -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:18:16.770 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:18:16.770 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:18:16.770 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:18:16.770 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:18:16.770 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:18:16.770 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:18:16.770 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:18:16.770 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:18:16.770 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.269 ms 00:18:16.770 00:18:16.770 --- 10.0.0.2 ping statistics --- 00:18:16.770 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:18:16.770 rtt min/avg/max/mdev = 0.269/0.269/0.269/0.000 ms 00:18:16.770 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:18:16.770 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:18:16.770 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.219 ms 00:18:16.770 00:18:16.770 --- 10.0.0.1 ping statistics --- 00:18:16.771 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:18:16.771 rtt min/avg/max/mdev = 0.219/0.219/0.219/0.000 ms 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@422 -- # return 0 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- target/tls.sh@63 -- # nvmfappstart -m 0x2 --wait-for-rpc 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@722 -- # xtrace_disable 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@481 -- # nvmfpid=612802 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@482 -- # waitforlisten 612802 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 --wait-for-rpc 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@829 -- # '[' -z 612802 ']' 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@834 -- # local max_retries=100 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:16.771 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@838 -- # xtrace_disable 00:18:16.771 18:04:10 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:18:16.771 [2024-07-15 18:04:10.217653] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:18:16.771 [2024-07-15 18:04:10.217697] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:18:16.771 EAL: No free 2048 kB hugepages reported on node 1 00:18:16.771 [2024-07-15 18:04:10.275262] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:16.771 [2024-07-15 18:04:10.353843] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:18:16.771 [2024-07-15 18:04:10.353875] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:18:16.771 [2024-07-15 18:04:10.353883] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:18:16.771 [2024-07-15 18:04:10.353889] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:18:16.771 [2024-07-15 18:04:10.353894] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:18:16.771 [2024-07-15 18:04:10.353917] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:18:17.339 18:04:11 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:18:17.339 18:04:11 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@862 -- # return 0 00:18:17.339 18:04:11 nvmf_tcp.nvmf_tls -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:18:17.339 18:04:11 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@728 -- # xtrace_disable 00:18:17.339 18:04:11 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:18:17.339 18:04:11 nvmf_tcp.nvmf_tls -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:18:17.339 18:04:11 nvmf_tcp.nvmf_tls -- target/tls.sh@65 -- # '[' tcp '!=' tcp ']' 00:18:17.339 18:04:11 nvmf_tcp.nvmf_tls -- target/tls.sh@70 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_set_default_impl -i ssl 00:18:17.598 true 00:18:17.598 18:04:11 nvmf_tcp.nvmf_tls -- target/tls.sh@73 -- # jq -r .tls_version 00:18:17.598 18:04:11 nvmf_tcp.nvmf_tls -- target/tls.sh@73 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:18:17.856 18:04:11 nvmf_tcp.nvmf_tls -- target/tls.sh@73 -- # version=0 00:18:17.856 18:04:11 nvmf_tcp.nvmf_tls -- target/tls.sh@74 -- # [[ 0 != \0 ]] 00:18:17.856 18:04:11 nvmf_tcp.nvmf_tls -- target/tls.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --tls-version 13 00:18:17.856 18:04:11 nvmf_tcp.nvmf_tls -- target/tls.sh@81 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:18:17.856 18:04:11 nvmf_tcp.nvmf_tls -- target/tls.sh@81 -- # jq -r .tls_version 00:18:18.114 18:04:11 nvmf_tcp.nvmf_tls -- target/tls.sh@81 -- # version=13 00:18:18.114 18:04:11 nvmf_tcp.nvmf_tls -- target/tls.sh@82 -- # [[ 13 != \1\3 ]] 00:18:18.114 18:04:11 nvmf_tcp.nvmf_tls -- target/tls.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --tls-version 7 00:18:18.375 18:04:11 nvmf_tcp.nvmf_tls -- target/tls.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:18:18.375 18:04:11 nvmf_tcp.nvmf_tls -- target/tls.sh@89 -- # jq -r .tls_version 00:18:18.375 18:04:12 nvmf_tcp.nvmf_tls -- target/tls.sh@89 -- # version=7 00:18:18.375 18:04:12 nvmf_tcp.nvmf_tls -- target/tls.sh@90 -- # [[ 7 != \7 ]] 00:18:18.375 18:04:12 nvmf_tcp.nvmf_tls -- target/tls.sh@96 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:18:18.375 18:04:12 nvmf_tcp.nvmf_tls -- target/tls.sh@96 -- # jq -r .enable_ktls 00:18:18.633 18:04:12 nvmf_tcp.nvmf_tls -- target/tls.sh@96 -- # ktls=false 00:18:18.633 18:04:12 nvmf_tcp.nvmf_tls -- target/tls.sh@97 -- # [[ false != \f\a\l\s\e ]] 00:18:18.633 18:04:12 nvmf_tcp.nvmf_tls -- target/tls.sh@103 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --enable-ktls 00:18:18.892 18:04:12 nvmf_tcp.nvmf_tls -- target/tls.sh@104 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:18:18.892 18:04:12 nvmf_tcp.nvmf_tls -- target/tls.sh@104 -- # jq -r .enable_ktls 00:18:18.892 18:04:12 nvmf_tcp.nvmf_tls -- target/tls.sh@104 -- # ktls=true 00:18:18.892 18:04:12 nvmf_tcp.nvmf_tls -- target/tls.sh@105 -- # [[ true != \t\r\u\e ]] 00:18:18.892 18:04:12 nvmf_tcp.nvmf_tls -- target/tls.sh@111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --disable-ktls 00:18:19.151 18:04:12 nvmf_tcp.nvmf_tls -- target/tls.sh@112 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_get_options -i ssl 00:18:19.151 18:04:12 nvmf_tcp.nvmf_tls -- target/tls.sh@112 -- # jq -r .enable_ktls 00:18:19.410 18:04:12 nvmf_tcp.nvmf_tls -- target/tls.sh@112 -- # ktls=false 00:18:19.410 18:04:12 nvmf_tcp.nvmf_tls -- target/tls.sh@113 -- # [[ false != \f\a\l\s\e ]] 00:18:19.411 18:04:12 nvmf_tcp.nvmf_tls -- target/tls.sh@118 -- # format_interchange_psk 00112233445566778899aabbccddeeff 1 00:18:19.411 18:04:12 nvmf_tcp.nvmf_tls -- nvmf/common.sh@715 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 1 00:18:19.411 18:04:12 nvmf_tcp.nvmf_tls -- nvmf/common.sh@702 -- # local prefix key digest 00:18:19.411 18:04:12 nvmf_tcp.nvmf_tls -- nvmf/common.sh@704 -- # prefix=NVMeTLSkey-1 00:18:19.411 18:04:12 nvmf_tcp.nvmf_tls -- nvmf/common.sh@704 -- # key=00112233445566778899aabbccddeeff 00:18:19.411 18:04:12 nvmf_tcp.nvmf_tls -- nvmf/common.sh@704 -- # digest=1 00:18:19.411 18:04:12 nvmf_tcp.nvmf_tls -- nvmf/common.sh@705 -- # python - 00:18:19.411 18:04:12 nvmf_tcp.nvmf_tls -- target/tls.sh@118 -- # key=NVMeTLSkey-1:01:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: 00:18:19.411 18:04:12 nvmf_tcp.nvmf_tls -- target/tls.sh@119 -- # format_interchange_psk ffeeddccbbaa99887766554433221100 1 00:18:19.411 18:04:12 nvmf_tcp.nvmf_tls -- nvmf/common.sh@715 -- # format_key NVMeTLSkey-1 ffeeddccbbaa99887766554433221100 1 00:18:19.411 18:04:12 nvmf_tcp.nvmf_tls -- nvmf/common.sh@702 -- # local prefix key digest 00:18:19.411 18:04:12 nvmf_tcp.nvmf_tls -- nvmf/common.sh@704 -- # prefix=NVMeTLSkey-1 00:18:19.411 18:04:12 nvmf_tcp.nvmf_tls -- nvmf/common.sh@704 -- # key=ffeeddccbbaa99887766554433221100 00:18:19.411 18:04:12 nvmf_tcp.nvmf_tls -- nvmf/common.sh@704 -- # digest=1 00:18:19.411 18:04:12 nvmf_tcp.nvmf_tls -- nvmf/common.sh@705 -- # python - 00:18:19.411 18:04:13 nvmf_tcp.nvmf_tls -- target/tls.sh@119 -- # key_2=NVMeTLSkey-1:01:ZmZlZWRkY2NiYmFhOTk4ODc3NjY1NTQ0MzMyMjExMDBfBm/Y: 00:18:19.411 18:04:13 nvmf_tcp.nvmf_tls -- target/tls.sh@121 -- # mktemp 00:18:19.411 18:04:13 nvmf_tcp.nvmf_tls -- target/tls.sh@121 -- # key_path=/tmp/tmp.q1a1GxY6ef 00:18:19.411 18:04:13 nvmf_tcp.nvmf_tls -- target/tls.sh@122 -- # mktemp 00:18:19.411 18:04:13 nvmf_tcp.nvmf_tls -- target/tls.sh@122 -- # key_2_path=/tmp/tmp.dBed4Y1irx 00:18:19.411 18:04:13 nvmf_tcp.nvmf_tls -- target/tls.sh@124 -- # echo -n NVMeTLSkey-1:01:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: 00:18:19.411 18:04:13 nvmf_tcp.nvmf_tls -- target/tls.sh@125 -- # echo -n NVMeTLSkey-1:01:ZmZlZWRkY2NiYmFhOTk4ODc3NjY1NTQ0MzMyMjExMDBfBm/Y: 00:18:19.411 18:04:13 nvmf_tcp.nvmf_tls -- target/tls.sh@127 -- # chmod 0600 /tmp/tmp.q1a1GxY6ef 00:18:19.411 18:04:13 nvmf_tcp.nvmf_tls -- target/tls.sh@128 -- # chmod 0600 /tmp/tmp.dBed4Y1irx 00:18:19.411 18:04:13 nvmf_tcp.nvmf_tls -- target/tls.sh@130 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py sock_impl_set_options -i ssl --tls-version 13 00:18:19.679 18:04:13 nvmf_tcp.nvmf_tls -- target/tls.sh@131 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py framework_start_init 00:18:19.939 18:04:13 nvmf_tcp.nvmf_tls -- target/tls.sh@133 -- # setup_nvmf_tgt /tmp/tmp.q1a1GxY6ef 00:18:19.939 18:04:13 nvmf_tcp.nvmf_tls -- target/tls.sh@49 -- # local key=/tmp/tmp.q1a1GxY6ef 00:18:19.939 18:04:13 nvmf_tcp.nvmf_tls -- target/tls.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:18:19.939 [2024-07-15 18:04:13.590228] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:18:19.939 18:04:13 nvmf_tcp.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:18:20.197 18:04:13 nvmf_tcp.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:18:20.198 [2024-07-15 18:04:13.923086] tcp.c: 966:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:18:20.198 [2024-07-15 18:04:13.923294] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:18:20.456 18:04:13 nvmf_tcp.nvmf_tls -- target/tls.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:18:20.456 malloc0 00:18:20.456 18:04:14 nvmf_tcp.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:18:20.714 18:04:14 nvmf_tcp.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.q1a1GxY6ef 00:18:20.714 [2024-07-15 18:04:14.420478] tcp.c:3710:nvmf_tcp_subsystem_add_host: *WARNING*: nvmf_tcp_psk_path: deprecated feature PSK path to be removed in v24.09 00:18:20.714 18:04:14 nvmf_tcp.nvmf_tls -- target/tls.sh@137 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -S ssl -q 64 -o 4096 -w randrw -M 30 -t 10 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 hostnqn:nqn.2016-06.io.spdk:host1' --psk-path /tmp/tmp.q1a1GxY6ef 00:18:20.973 EAL: No free 2048 kB hugepages reported on node 1 00:18:30.951 Initializing NVMe Controllers 00:18:30.951 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:18:30.951 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:18:30.951 Initialization complete. Launching workers. 00:18:30.951 ======================================================== 00:18:30.951 Latency(us) 00:18:30.951 Device Information : IOPS MiB/s Average min max 00:18:30.951 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 16211.76 63.33 3948.19 742.76 205746.01 00:18:30.951 ======================================================== 00:18:30.951 Total : 16211.76 63.33 3948.19 742.76 205746.01 00:18:30.951 00:18:30.951 18:04:24 nvmf_tcp.nvmf_tls -- target/tls.sh@143 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.q1a1GxY6ef 00:18:30.951 18:04:24 nvmf_tcp.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:18:30.951 18:04:24 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:18:30.951 18:04:24 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:18:30.951 18:04:24 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # psk='--psk /tmp/tmp.q1a1GxY6ef' 00:18:30.951 18:04:24 nvmf_tcp.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:18:30.951 18:04:24 nvmf_tcp.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=615152 00:18:30.951 18:04:24 nvmf_tcp.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:18:30.951 18:04:24 nvmf_tcp.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:18:30.951 18:04:24 nvmf_tcp.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 615152 /var/tmp/bdevperf.sock 00:18:30.951 18:04:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@829 -- # '[' -z 615152 ']' 00:18:30.951 18:04:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:18:30.951 18:04:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@834 -- # local max_retries=100 00:18:30.951 18:04:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:18:30.951 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:18:30.951 18:04:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@838 -- # xtrace_disable 00:18:30.951 18:04:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:18:30.951 [2024-07-15 18:04:24.577185] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:18:30.951 [2024-07-15 18:04:24.577243] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid615152 ] 00:18:30.951 EAL: No free 2048 kB hugepages reported on node 1 00:18:30.951 [2024-07-15 18:04:24.625568] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:31.251 [2024-07-15 18:04:24.698137] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:18:31.846 18:04:25 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:18:31.846 18:04:25 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@862 -- # return 0 00:18:31.846 18:04:25 nvmf_tcp.nvmf_tls -- target/tls.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.q1a1GxY6ef 00:18:31.846 [2024-07-15 18:04:25.536666] bdev_nvme_rpc.c: 517:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:18:31.846 [2024-07-15 18:04:25.536729] nvme_tcp.c:2589:nvme_tcp_generate_tls_credentials: *WARNING*: nvme_ctrlr_psk: deprecated feature spdk_nvme_ctrlr_opts.psk to be removed in v24.09 00:18:32.104 TLSTESTn1 00:18:32.104 18:04:25 nvmf_tcp.nvmf_tls -- target/tls.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 20 -s /var/tmp/bdevperf.sock perform_tests 00:18:32.104 Running I/O for 10 seconds... 00:18:42.089 00:18:42.089 Latency(us) 00:18:42.089 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:18:42.089 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:18:42.089 Verification LBA range: start 0x0 length 0x2000 00:18:42.089 TLSTESTn1 : 10.05 2830.27 11.06 0.00 0.00 45109.29 7151.97 53112.65 00:18:42.089 =================================================================================================================== 00:18:42.090 Total : 2830.27 11.06 0.00 0.00 45109.29 7151.97 53112.65 00:18:42.090 0 00:18:42.090 18:04:35 nvmf_tcp.nvmf_tls -- target/tls.sh@44 -- # trap 'nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:18:42.090 18:04:35 nvmf_tcp.nvmf_tls -- target/tls.sh@45 -- # killprocess 615152 00:18:42.090 18:04:35 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@948 -- # '[' -z 615152 ']' 00:18:42.090 18:04:35 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@952 -- # kill -0 615152 00:18:42.349 18:04:35 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # uname 00:18:42.349 18:04:35 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:18:42.349 18:04:35 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 615152 00:18:42.349 18:04:35 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # process_name=reactor_2 00:18:42.349 18:04:35 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@958 -- # '[' reactor_2 = sudo ']' 00:18:42.349 18:04:35 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@966 -- # echo 'killing process with pid 615152' 00:18:42.349 killing process with pid 615152 00:18:42.349 18:04:35 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@967 -- # kill 615152 00:18:42.349 Received shutdown signal, test time was about 10.000000 seconds 00:18:42.349 00:18:42.349 Latency(us) 00:18:42.349 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:18:42.349 =================================================================================================================== 00:18:42.349 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:18:42.349 [2024-07-15 18:04:35.860940] app.c:1024:log_deprecation_hits: *WARNING*: nvme_ctrlr_psk: deprecation 'spdk_nvme_ctrlr_opts.psk' scheduled for removal in v24.09 hit 1 times 00:18:42.349 18:04:35 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@972 -- # wait 615152 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- target/tls.sh@146 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.dBed4Y1irx 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@648 -- # local es=0 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@650 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.dBed4Y1irx 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@636 -- # local arg=run_bdevperf 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@640 -- # type -t run_bdevperf 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@651 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.dBed4Y1irx 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # psk='--psk /tmp/tmp.dBed4Y1irx' 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=617053 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 617053 /var/tmp/bdevperf.sock 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@829 -- # '[' -z 617053 ']' 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@834 -- # local max_retries=100 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:18:42.349 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@838 -- # xtrace_disable 00:18:42.349 18:04:36 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:18:42.609 [2024-07-15 18:04:36.090718] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:18:42.609 [2024-07-15 18:04:36.090765] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid617053 ] 00:18:42.609 EAL: No free 2048 kB hugepages reported on node 1 00:18:42.609 [2024-07-15 18:04:36.140902] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:42.609 [2024-07-15 18:04:36.222180] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:18:43.544 18:04:36 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:18:43.544 18:04:36 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@862 -- # return 0 00:18:43.544 18:04:36 nvmf_tcp.nvmf_tls -- target/tls.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.dBed4Y1irx 00:18:43.544 [2024-07-15 18:04:37.049193] bdev_nvme_rpc.c: 517:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:18:43.544 [2024-07-15 18:04:37.049273] nvme_tcp.c:2589:nvme_tcp_generate_tls_credentials: *WARNING*: nvme_ctrlr_psk: deprecated feature spdk_nvme_ctrlr_opts.psk to be removed in v24.09 00:18:43.544 [2024-07-15 18:04:37.057605] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 428:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:18:43.544 [2024-07-15 18:04:37.058467] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x23f6570 (107): Transport endpoint is not connected 00:18:43.544 [2024-07-15 18:04:37.059460] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x23f6570 (9): Bad file descriptor 00:18:43.544 [2024-07-15 18:04:37.060461] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:18:43.544 [2024-07-15 18:04:37.060470] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 10.0.0.2 00:18:43.544 [2024-07-15 18:04:37.060478] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:18:43.544 request: 00:18:43.544 { 00:18:43.544 "name": "TLSTEST", 00:18:43.544 "trtype": "tcp", 00:18:43.544 "traddr": "10.0.0.2", 00:18:43.544 "adrfam": "ipv4", 00:18:43.544 "trsvcid": "4420", 00:18:43.544 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:18:43.544 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:18:43.544 "prchk_reftag": false, 00:18:43.544 "prchk_guard": false, 00:18:43.544 "hdgst": false, 00:18:43.544 "ddgst": false, 00:18:43.544 "psk": "/tmp/tmp.dBed4Y1irx", 00:18:43.544 "method": "bdev_nvme_attach_controller", 00:18:43.544 "req_id": 1 00:18:43.544 } 00:18:43.544 Got JSON-RPC error response 00:18:43.544 response: 00:18:43.544 { 00:18:43.544 "code": -5, 00:18:43.544 "message": "Input/output error" 00:18:43.544 } 00:18:43.544 18:04:37 nvmf_tcp.nvmf_tls -- target/tls.sh@36 -- # killprocess 617053 00:18:43.544 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@948 -- # '[' -z 617053 ']' 00:18:43.544 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@952 -- # kill -0 617053 00:18:43.544 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # uname 00:18:43.544 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:18:43.544 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 617053 00:18:43.544 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # process_name=reactor_2 00:18:43.544 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@958 -- # '[' reactor_2 = sudo ']' 00:18:43.544 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@966 -- # echo 'killing process with pid 617053' 00:18:43.544 killing process with pid 617053 00:18:43.544 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@967 -- # kill 617053 00:18:43.544 Received shutdown signal, test time was about 10.000000 seconds 00:18:43.544 00:18:43.544 Latency(us) 00:18:43.544 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:18:43.544 =================================================================================================================== 00:18:43.544 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:18:43.544 [2024-07-15 18:04:37.119764] app.c:1024:log_deprecation_hits: *WARNING*: nvme_ctrlr_psk: deprecation 'spdk_nvme_ctrlr_opts.psk' scheduled for removal in v24.09 hit 1 times 00:18:43.544 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@972 -- # wait 617053 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- target/tls.sh@37 -- # return 1 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@651 -- # es=1 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- target/tls.sh@149 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host2 /tmp/tmp.q1a1GxY6ef 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@648 -- # local es=0 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@650 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host2 /tmp/tmp.q1a1GxY6ef 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@636 -- # local arg=run_bdevperf 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@640 -- # type -t run_bdevperf 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@651 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host2 /tmp/tmp.q1a1GxY6ef 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host2 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # psk='--psk /tmp/tmp.q1a1GxY6ef' 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=617226 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 617226 /var/tmp/bdevperf.sock 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@829 -- # '[' -z 617226 ']' 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@834 -- # local max_retries=100 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:18:43.803 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@838 -- # xtrace_disable 00:18:43.803 18:04:37 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:18:43.803 [2024-07-15 18:04:37.339024] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:18:43.803 [2024-07-15 18:04:37.339072] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid617226 ] 00:18:43.803 EAL: No free 2048 kB hugepages reported on node 1 00:18:43.803 [2024-07-15 18:04:37.389572] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:43.803 [2024-07-15 18:04:37.468359] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:18:44.737 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:18:44.737 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@862 -- # return 0 00:18:44.737 18:04:38 nvmf_tcp.nvmf_tls -- target/tls.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host2 --psk /tmp/tmp.q1a1GxY6ef 00:18:44.737 [2024-07-15 18:04:38.302243] bdev_nvme_rpc.c: 517:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:18:44.737 [2024-07-15 18:04:38.302305] nvme_tcp.c:2589:nvme_tcp_generate_tls_credentials: *WARNING*: nvme_ctrlr_psk: deprecated feature spdk_nvme_ctrlr_opts.psk to be removed in v24.09 00:18:44.737 [2024-07-15 18:04:38.310699] tcp.c: 918:tcp_sock_get_key: *ERROR*: Could not find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host2 nqn.2016-06.io.spdk:cnode1 00:18:44.737 [2024-07-15 18:04:38.310720] posix.c: 589:posix_sock_psk_find_session_server_cb: *ERROR*: Unable to find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host2 nqn.2016-06.io.spdk:cnode1 00:18:44.738 [2024-07-15 18:04:38.310744] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 428:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:18:44.738 [2024-07-15 18:04:38.311518] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xd4e570 (107): Transport endpoint is not connected 00:18:44.738 [2024-07-15 18:04:38.312512] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0xd4e570 (9): Bad file descriptor 00:18:44.738 [2024-07-15 18:04:38.313513] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:18:44.738 [2024-07-15 18:04:38.313522] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 10.0.0.2 00:18:44.738 [2024-07-15 18:04:38.313530] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:18:44.738 request: 00:18:44.738 { 00:18:44.738 "name": "TLSTEST", 00:18:44.738 "trtype": "tcp", 00:18:44.738 "traddr": "10.0.0.2", 00:18:44.738 "adrfam": "ipv4", 00:18:44.738 "trsvcid": "4420", 00:18:44.738 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:18:44.738 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:18:44.738 "prchk_reftag": false, 00:18:44.738 "prchk_guard": false, 00:18:44.738 "hdgst": false, 00:18:44.738 "ddgst": false, 00:18:44.738 "psk": "/tmp/tmp.q1a1GxY6ef", 00:18:44.738 "method": "bdev_nvme_attach_controller", 00:18:44.738 "req_id": 1 00:18:44.738 } 00:18:44.738 Got JSON-RPC error response 00:18:44.738 response: 00:18:44.738 { 00:18:44.738 "code": -5, 00:18:44.738 "message": "Input/output error" 00:18:44.738 } 00:18:44.738 18:04:38 nvmf_tcp.nvmf_tls -- target/tls.sh@36 -- # killprocess 617226 00:18:44.738 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@948 -- # '[' -z 617226 ']' 00:18:44.738 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@952 -- # kill -0 617226 00:18:44.738 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # uname 00:18:44.738 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:18:44.738 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 617226 00:18:44.738 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # process_name=reactor_2 00:18:44.738 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@958 -- # '[' reactor_2 = sudo ']' 00:18:44.738 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@966 -- # echo 'killing process with pid 617226' 00:18:44.738 killing process with pid 617226 00:18:44.738 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@967 -- # kill 617226 00:18:44.738 Received shutdown signal, test time was about 10.000000 seconds 00:18:44.738 00:18:44.738 Latency(us) 00:18:44.738 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:18:44.738 =================================================================================================================== 00:18:44.738 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:18:44.738 [2024-07-15 18:04:38.389614] app.c:1024:log_deprecation_hits: *WARNING*: nvme_ctrlr_psk: deprecation 'spdk_nvme_ctrlr_opts.psk' scheduled for removal in v24.09 hit 1 times 00:18:44.738 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@972 -- # wait 617226 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- target/tls.sh@37 -- # return 1 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@651 -- # es=1 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- target/tls.sh@152 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode2 nqn.2016-06.io.spdk:host1 /tmp/tmp.q1a1GxY6ef 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@648 -- # local es=0 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@650 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode2 nqn.2016-06.io.spdk:host1 /tmp/tmp.q1a1GxY6ef 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@636 -- # local arg=run_bdevperf 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@640 -- # type -t run_bdevperf 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@651 -- # run_bdevperf nqn.2016-06.io.spdk:cnode2 nqn.2016-06.io.spdk:host1 /tmp/tmp.q1a1GxY6ef 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode2 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # psk='--psk /tmp/tmp.q1a1GxY6ef' 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=617470 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 617470 /var/tmp/bdevperf.sock 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@829 -- # '[' -z 617470 ']' 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@834 -- # local max_retries=100 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:18:44.996 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@838 -- # xtrace_disable 00:18:44.996 18:04:38 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:18:44.996 [2024-07-15 18:04:38.617341] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:18:44.996 [2024-07-15 18:04:38.617389] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid617470 ] 00:18:44.996 EAL: No free 2048 kB hugepages reported on node 1 00:18:44.996 [2024-07-15 18:04:38.667369] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:45.254 [2024-07-15 18:04:38.737570] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:18:45.820 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:18:45.820 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@862 -- # return 0 00:18:45.820 18:04:39 nvmf_tcp.nvmf_tls -- target/tls.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -q nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.q1a1GxY6ef 00:18:46.079 [2024-07-15 18:04:39.571856] bdev_nvme_rpc.c: 517:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:18:46.079 [2024-07-15 18:04:39.571933] nvme_tcp.c:2589:nvme_tcp_generate_tls_credentials: *WARNING*: nvme_ctrlr_psk: deprecated feature spdk_nvme_ctrlr_opts.psk to be removed in v24.09 00:18:46.079 [2024-07-15 18:04:39.576418] tcp.c: 918:tcp_sock_get_key: *ERROR*: Could not find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host1 nqn.2016-06.io.spdk:cnode2 00:18:46.079 [2024-07-15 18:04:39.576438] posix.c: 589:posix_sock_psk_find_session_server_cb: *ERROR*: Unable to find PSK for identity: NVMe0R01 nqn.2016-06.io.spdk:host1 nqn.2016-06.io.spdk:cnode2 00:18:46.079 [2024-07-15 18:04:39.576461] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 428:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:18:46.079 [2024-07-15 18:04:39.577140] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1f62570 (107): Transport endpoint is not connected 00:18:46.079 [2024-07-15 18:04:39.578132] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1f62570 (9): Bad file descriptor 00:18:46.079 [2024-07-15 18:04:39.579132] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode2] Ctrlr is in error state 00:18:46.079 [2024-07-15 18:04:39.579141] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 10.0.0.2 00:18:46.079 [2024-07-15 18:04:39.579152] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode2] in failed state. 00:18:46.079 request: 00:18:46.079 { 00:18:46.079 "name": "TLSTEST", 00:18:46.079 "trtype": "tcp", 00:18:46.079 "traddr": "10.0.0.2", 00:18:46.079 "adrfam": "ipv4", 00:18:46.079 "trsvcid": "4420", 00:18:46.079 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:18:46.079 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:18:46.079 "prchk_reftag": false, 00:18:46.079 "prchk_guard": false, 00:18:46.079 "hdgst": false, 00:18:46.079 "ddgst": false, 00:18:46.079 "psk": "/tmp/tmp.q1a1GxY6ef", 00:18:46.079 "method": "bdev_nvme_attach_controller", 00:18:46.079 "req_id": 1 00:18:46.079 } 00:18:46.079 Got JSON-RPC error response 00:18:46.079 response: 00:18:46.079 { 00:18:46.079 "code": -5, 00:18:46.079 "message": "Input/output error" 00:18:46.079 } 00:18:46.079 18:04:39 nvmf_tcp.nvmf_tls -- target/tls.sh@36 -- # killprocess 617470 00:18:46.079 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@948 -- # '[' -z 617470 ']' 00:18:46.079 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@952 -- # kill -0 617470 00:18:46.079 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # uname 00:18:46.079 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:18:46.079 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 617470 00:18:46.079 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # process_name=reactor_2 00:18:46.079 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@958 -- # '[' reactor_2 = sudo ']' 00:18:46.079 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@966 -- # echo 'killing process with pid 617470' 00:18:46.079 killing process with pid 617470 00:18:46.079 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@967 -- # kill 617470 00:18:46.079 Received shutdown signal, test time was about 10.000000 seconds 00:18:46.079 00:18:46.079 Latency(us) 00:18:46.079 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:18:46.079 =================================================================================================================== 00:18:46.079 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:18:46.079 [2024-07-15 18:04:39.643541] app.c:1024:log_deprecation_hits: *WARNING*: nvme_ctrlr_psk: deprecation 'spdk_nvme_ctrlr_opts.psk' scheduled for removal in v24.09 hit 1 times 00:18:46.079 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@972 -- # wait 617470 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- target/tls.sh@37 -- # return 1 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@651 -- # es=1 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- target/tls.sh@155 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 '' 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@648 -- # local es=0 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@650 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 '' 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@636 -- # local arg=run_bdevperf 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@640 -- # type -t run_bdevperf 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@651 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 '' 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # psk= 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=617704 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 617704 /var/tmp/bdevperf.sock 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@829 -- # '[' -z 617704 ']' 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@834 -- # local max_retries=100 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:18:46.338 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@838 -- # xtrace_disable 00:18:46.338 18:04:39 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:18:46.338 [2024-07-15 18:04:39.864135] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:18:46.338 [2024-07-15 18:04:39.864182] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid617704 ] 00:18:46.338 EAL: No free 2048 kB hugepages reported on node 1 00:18:46.338 [2024-07-15 18:04:39.913727] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:46.338 [2024-07-15 18:04:39.982938] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:18:47.273 18:04:40 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:18:47.273 18:04:40 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@862 -- # return 0 00:18:47.273 18:04:40 nvmf_tcp.nvmf_tls -- target/tls.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 00:18:47.273 [2024-07-15 18:04:40.837159] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 428:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:18:47.273 [2024-07-15 18:04:40.838800] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1456af0 (9): Bad file descriptor 00:18:47.273 [2024-07-15 18:04:40.839798] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:18:47.273 [2024-07-15 18:04:40.839808] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 10.0.0.2 00:18:47.273 [2024-07-15 18:04:40.839816] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:18:47.273 request: 00:18:47.273 { 00:18:47.273 "name": "TLSTEST", 00:18:47.273 "trtype": "tcp", 00:18:47.273 "traddr": "10.0.0.2", 00:18:47.273 "adrfam": "ipv4", 00:18:47.273 "trsvcid": "4420", 00:18:47.273 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:18:47.273 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:18:47.273 "prchk_reftag": false, 00:18:47.273 "prchk_guard": false, 00:18:47.273 "hdgst": false, 00:18:47.273 "ddgst": false, 00:18:47.273 "method": "bdev_nvme_attach_controller", 00:18:47.273 "req_id": 1 00:18:47.273 } 00:18:47.273 Got JSON-RPC error response 00:18:47.273 response: 00:18:47.273 { 00:18:47.273 "code": -5, 00:18:47.273 "message": "Input/output error" 00:18:47.273 } 00:18:47.273 18:04:40 nvmf_tcp.nvmf_tls -- target/tls.sh@36 -- # killprocess 617704 00:18:47.273 18:04:40 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@948 -- # '[' -z 617704 ']' 00:18:47.273 18:04:40 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@952 -- # kill -0 617704 00:18:47.273 18:04:40 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # uname 00:18:47.273 18:04:40 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:18:47.273 18:04:40 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 617704 00:18:47.273 18:04:40 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # process_name=reactor_2 00:18:47.273 18:04:40 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@958 -- # '[' reactor_2 = sudo ']' 00:18:47.274 18:04:40 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@966 -- # echo 'killing process with pid 617704' 00:18:47.274 killing process with pid 617704 00:18:47.274 18:04:40 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@967 -- # kill 617704 00:18:47.274 Received shutdown signal, test time was about 10.000000 seconds 00:18:47.274 00:18:47.274 Latency(us) 00:18:47.274 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:18:47.274 =================================================================================================================== 00:18:47.274 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:18:47.274 18:04:40 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@972 -- # wait 617704 00:18:47.532 18:04:41 nvmf_tcp.nvmf_tls -- target/tls.sh@37 -- # return 1 00:18:47.532 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@651 -- # es=1 00:18:47.532 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:18:47.532 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:18:47.532 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:18:47.532 18:04:41 nvmf_tcp.nvmf_tls -- target/tls.sh@158 -- # killprocess 612802 00:18:47.532 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@948 -- # '[' -z 612802 ']' 00:18:47.532 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@952 -- # kill -0 612802 00:18:47.532 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # uname 00:18:47.532 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:18:47.532 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 612802 00:18:47.532 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:18:47.532 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:18:47.532 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@966 -- # echo 'killing process with pid 612802' 00:18:47.532 killing process with pid 612802 00:18:47.532 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@967 -- # kill 612802 00:18:47.532 [2024-07-15 18:04:41.119405] app.c:1024:log_deprecation_hits: *WARNING*: nvmf_tcp_psk_path: deprecation 'PSK path' scheduled for removal in v24.09 hit 1 times 00:18:47.532 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@972 -- # wait 612802 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- target/tls.sh@159 -- # format_interchange_psk 00112233445566778899aabbccddeeff0011223344556677 2 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- nvmf/common.sh@715 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff0011223344556677 2 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- nvmf/common.sh@702 -- # local prefix key digest 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- nvmf/common.sh@704 -- # prefix=NVMeTLSkey-1 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- nvmf/common.sh@704 -- # key=00112233445566778899aabbccddeeff0011223344556677 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- nvmf/common.sh@704 -- # digest=2 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- nvmf/common.sh@705 -- # python - 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- target/tls.sh@159 -- # key_long=NVMeTLSkey-1:02:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmYwMDExMjIzMzQ0NTU2Njc3wWXNJw==: 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- target/tls.sh@160 -- # mktemp 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- target/tls.sh@160 -- # key_long_path=/tmp/tmp.BYHgc4Ja0k 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- target/tls.sh@161 -- # echo -n NVMeTLSkey-1:02:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmYwMDExMjIzMzQ0NTU2Njc3wWXNJw==: 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- target/tls.sh@162 -- # chmod 0600 /tmp/tmp.BYHgc4Ja0k 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- target/tls.sh@163 -- # nvmfappstart -m 0x2 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@722 -- # xtrace_disable 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- nvmf/common.sh@481 -- # nvmfpid=617954 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- nvmf/common.sh@482 -- # waitforlisten 617954 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@829 -- # '[' -z 617954 ']' 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@834 -- # local max_retries=100 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:18:47.792 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@838 -- # xtrace_disable 00:18:47.792 18:04:41 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:18:47.792 [2024-07-15 18:04:41.409739] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:18:47.792 [2024-07-15 18:04:41.409785] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:18:47.792 EAL: No free 2048 kB hugepages reported on node 1 00:18:47.792 [2024-07-15 18:04:41.465878] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:48.051 [2024-07-15 18:04:41.532796] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:18:48.051 [2024-07-15 18:04:41.532835] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:18:48.051 [2024-07-15 18:04:41.532843] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:18:48.051 [2024-07-15 18:04:41.532849] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:18:48.051 [2024-07-15 18:04:41.532853] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:18:48.051 [2024-07-15 18:04:41.532876] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:18:48.618 18:04:42 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:18:48.618 18:04:42 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@862 -- # return 0 00:18:48.618 18:04:42 nvmf_tcp.nvmf_tls -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:18:48.618 18:04:42 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@728 -- # xtrace_disable 00:18:48.618 18:04:42 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:18:48.618 18:04:42 nvmf_tcp.nvmf_tls -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:18:48.618 18:04:42 nvmf_tcp.nvmf_tls -- target/tls.sh@165 -- # setup_nvmf_tgt /tmp/tmp.BYHgc4Ja0k 00:18:48.618 18:04:42 nvmf_tcp.nvmf_tls -- target/tls.sh@49 -- # local key=/tmp/tmp.BYHgc4Ja0k 00:18:48.618 18:04:42 nvmf_tcp.nvmf_tls -- target/tls.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:18:48.876 [2024-07-15 18:04:42.392113] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:18:48.876 18:04:42 nvmf_tcp.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:18:48.876 18:04:42 nvmf_tcp.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:18:49.135 [2024-07-15 18:04:42.716928] tcp.c: 966:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:18:49.135 [2024-07-15 18:04:42.717122] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:18:49.135 18:04:42 nvmf_tcp.nvmf_tls -- target/tls.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:18:49.394 malloc0 00:18:49.395 18:04:42 nvmf_tcp.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:18:49.395 18:04:43 nvmf_tcp.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.BYHgc4Ja0k 00:18:49.654 [2024-07-15 18:04:43.214255] tcp.c:3710:nvmf_tcp_subsystem_add_host: *WARNING*: nvmf_tcp_psk_path: deprecated feature PSK path to be removed in v24.09 00:18:49.654 18:04:43 nvmf_tcp.nvmf_tls -- target/tls.sh@167 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.BYHgc4Ja0k 00:18:49.654 18:04:43 nvmf_tcp.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:18:49.654 18:04:43 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:18:49.654 18:04:43 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:18:49.654 18:04:43 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # psk='--psk /tmp/tmp.BYHgc4Ja0k' 00:18:49.654 18:04:43 nvmf_tcp.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:18:49.654 18:04:43 nvmf_tcp.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:18:49.654 18:04:43 nvmf_tcp.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=618366 00:18:49.654 18:04:43 nvmf_tcp.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:18:49.654 18:04:43 nvmf_tcp.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 618366 /var/tmp/bdevperf.sock 00:18:49.654 18:04:43 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@829 -- # '[' -z 618366 ']' 00:18:49.654 18:04:43 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:18:49.654 18:04:43 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@834 -- # local max_retries=100 00:18:49.654 18:04:43 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:18:49.654 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:18:49.654 18:04:43 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@838 -- # xtrace_disable 00:18:49.654 18:04:43 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:18:49.654 [2024-07-15 18:04:43.258035] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:18:49.655 [2024-07-15 18:04:43.258081] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid618366 ] 00:18:49.655 EAL: No free 2048 kB hugepages reported on node 1 00:18:49.655 [2024-07-15 18:04:43.308233] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:18:49.914 [2024-07-15 18:04:43.387228] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:18:49.914 18:04:43 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:18:49.914 18:04:43 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@862 -- # return 0 00:18:49.914 18:04:43 nvmf_tcp.nvmf_tls -- target/tls.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.BYHgc4Ja0k 00:18:49.914 [2024-07-15 18:04:43.628359] bdev_nvme_rpc.c: 517:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:18:49.914 [2024-07-15 18:04:43.628431] nvme_tcp.c:2589:nvme_tcp_generate_tls_credentials: *WARNING*: nvme_ctrlr_psk: deprecated feature spdk_nvme_ctrlr_opts.psk to be removed in v24.09 00:18:50.173 TLSTESTn1 00:18:50.173 18:04:43 nvmf_tcp.nvmf_tls -- target/tls.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 20 -s /var/tmp/bdevperf.sock perform_tests 00:18:50.173 Running I/O for 10 seconds... 00:19:00.177 00:19:00.177 Latency(us) 00:19:00.177 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:00.177 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:19:00.177 Verification LBA range: start 0x0 length 0x2000 00:19:00.177 TLSTESTn1 : 10.03 4314.69 16.85 0.00 0.00 29614.23 4729.99 62914.56 00:19:00.177 =================================================================================================================== 00:19:00.177 Total : 4314.69 16.85 0.00 0.00 29614.23 4729.99 62914.56 00:19:00.177 0 00:19:00.177 18:04:53 nvmf_tcp.nvmf_tls -- target/tls.sh@44 -- # trap 'nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:19:00.177 18:04:53 nvmf_tcp.nvmf_tls -- target/tls.sh@45 -- # killprocess 618366 00:19:00.177 18:04:53 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@948 -- # '[' -z 618366 ']' 00:19:00.177 18:04:53 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@952 -- # kill -0 618366 00:19:00.177 18:04:53 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # uname 00:19:00.177 18:04:53 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:19:00.177 18:04:53 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 618366 00:19:00.437 18:04:53 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # process_name=reactor_2 00:19:00.437 18:04:53 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@958 -- # '[' reactor_2 = sudo ']' 00:19:00.437 18:04:53 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@966 -- # echo 'killing process with pid 618366' 00:19:00.437 killing process with pid 618366 00:19:00.437 18:04:53 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@967 -- # kill 618366 00:19:00.437 Received shutdown signal, test time was about 10.000000 seconds 00:19:00.437 00:19:00.437 Latency(us) 00:19:00.437 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:00.437 =================================================================================================================== 00:19:00.437 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:19:00.437 [2024-07-15 18:04:53.927966] app.c:1024:log_deprecation_hits: *WARNING*: nvme_ctrlr_psk: deprecation 'spdk_nvme_ctrlr_opts.psk' scheduled for removal in v24.09 hit 1 times 00:19:00.437 18:04:53 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@972 -- # wait 618366 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- target/tls.sh@170 -- # chmod 0666 /tmp/tmp.BYHgc4Ja0k 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- target/tls.sh@171 -- # NOT run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.BYHgc4Ja0k 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@648 -- # local es=0 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@650 -- # valid_exec_arg run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.BYHgc4Ja0k 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@636 -- # local arg=run_bdevperf 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@640 -- # type -t run_bdevperf 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@651 -- # run_bdevperf nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 /tmp/tmp.BYHgc4Ja0k 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- target/tls.sh@22 -- # local subnqn hostnqn psk 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # subnqn=nqn.2016-06.io.spdk:cnode1 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # hostnqn=nqn.2016-06.io.spdk:host1 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- target/tls.sh@23 -- # psk='--psk /tmp/tmp.BYHgc4Ja0k' 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- target/tls.sh@25 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- target/tls.sh@28 -- # bdevperf_pid=620049 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- target/tls.sh@30 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- target/tls.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- target/tls.sh@31 -- # waitforlisten 620049 /var/tmp/bdevperf.sock 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@829 -- # '[' -z 620049 ']' 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@834 -- # local max_retries=100 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:00.437 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@838 -- # xtrace_disable 00:19:00.437 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:00.437 [2024-07-15 18:04:54.157592] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:19:00.437 [2024-07-15 18:04:54.157640] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid620049 ] 00:19:00.696 EAL: No free 2048 kB hugepages reported on node 1 00:19:00.696 [2024-07-15 18:04:54.207689] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:00.696 [2024-07-15 18:04:54.277119] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:19:00.696 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:19:00.696 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@862 -- # return 0 00:19:00.696 18:04:54 nvmf_tcp.nvmf_tls -- target/tls.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.BYHgc4Ja0k 00:19:00.955 [2024-07-15 18:04:54.526343] bdev_nvme_rpc.c: 517:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:00.955 [2024-07-15 18:04:54.526393] bdev_nvme.c:6125:bdev_nvme_load_psk: *ERROR*: Incorrect permissions for PSK file 00:19:00.955 [2024-07-15 18:04:54.526400] bdev_nvme.c:6230:bdev_nvme_create: *ERROR*: Could not load PSK from /tmp/tmp.BYHgc4Ja0k 00:19:00.955 request: 00:19:00.955 { 00:19:00.955 "name": "TLSTEST", 00:19:00.955 "trtype": "tcp", 00:19:00.955 "traddr": "10.0.0.2", 00:19:00.955 "adrfam": "ipv4", 00:19:00.955 "trsvcid": "4420", 00:19:00.955 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:19:00.955 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:19:00.955 "prchk_reftag": false, 00:19:00.955 "prchk_guard": false, 00:19:00.955 "hdgst": false, 00:19:00.955 "ddgst": false, 00:19:00.955 "psk": "/tmp/tmp.BYHgc4Ja0k", 00:19:00.955 "method": "bdev_nvme_attach_controller", 00:19:00.955 "req_id": 1 00:19:00.955 } 00:19:00.955 Got JSON-RPC error response 00:19:00.955 response: 00:19:00.955 { 00:19:00.955 "code": -1, 00:19:00.955 "message": "Operation not permitted" 00:19:00.955 } 00:19:00.955 18:04:54 nvmf_tcp.nvmf_tls -- target/tls.sh@36 -- # killprocess 620049 00:19:00.955 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@948 -- # '[' -z 620049 ']' 00:19:00.955 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@952 -- # kill -0 620049 00:19:00.955 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # uname 00:19:00.955 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:19:00.955 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 620049 00:19:00.955 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # process_name=reactor_2 00:19:00.955 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@958 -- # '[' reactor_2 = sudo ']' 00:19:00.955 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@966 -- # echo 'killing process with pid 620049' 00:19:00.955 killing process with pid 620049 00:19:00.955 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@967 -- # kill 620049 00:19:00.955 Received shutdown signal, test time was about 10.000000 seconds 00:19:00.955 00:19:00.955 Latency(us) 00:19:00.955 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:00.955 =================================================================================================================== 00:19:00.955 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:19:00.955 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@972 -- # wait 620049 00:19:01.215 18:04:54 nvmf_tcp.nvmf_tls -- target/tls.sh@37 -- # return 1 00:19:01.215 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@651 -- # es=1 00:19:01.215 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:19:01.215 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:19:01.215 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:19:01.215 18:04:54 nvmf_tcp.nvmf_tls -- target/tls.sh@174 -- # killprocess 617954 00:19:01.215 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@948 -- # '[' -z 617954 ']' 00:19:01.215 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@952 -- # kill -0 617954 00:19:01.215 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # uname 00:19:01.215 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:19:01.215 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 617954 00:19:01.215 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:19:01.215 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:19:01.215 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@966 -- # echo 'killing process with pid 617954' 00:19:01.215 killing process with pid 617954 00:19:01.215 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@967 -- # kill 617954 00:19:01.215 [2024-07-15 18:04:54.812620] app.c:1024:log_deprecation_hits: *WARNING*: nvmf_tcp_psk_path: deprecation 'PSK path' scheduled for removal in v24.09 hit 1 times 00:19:01.215 18:04:54 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@972 -- # wait 617954 00:19:01.474 18:04:55 nvmf_tcp.nvmf_tls -- target/tls.sh@175 -- # nvmfappstart -m 0x2 00:19:01.474 18:04:55 nvmf_tcp.nvmf_tls -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:19:01.474 18:04:55 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@722 -- # xtrace_disable 00:19:01.474 18:04:55 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:01.474 18:04:55 nvmf_tcp.nvmf_tls -- nvmf/common.sh@481 -- # nvmfpid=620283 00:19:01.474 18:04:55 nvmf_tcp.nvmf_tls -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:19:01.474 18:04:55 nvmf_tcp.nvmf_tls -- nvmf/common.sh@482 -- # waitforlisten 620283 00:19:01.474 18:04:55 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@829 -- # '[' -z 620283 ']' 00:19:01.474 18:04:55 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:01.474 18:04:55 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@834 -- # local max_retries=100 00:19:01.474 18:04:55 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:01.474 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:01.474 18:04:55 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@838 -- # xtrace_disable 00:19:01.474 18:04:55 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:01.474 [2024-07-15 18:04:55.058711] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:19:01.474 [2024-07-15 18:04:55.058759] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:01.474 EAL: No free 2048 kB hugepages reported on node 1 00:19:01.474 [2024-07-15 18:04:55.113696] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:01.474 [2024-07-15 18:04:55.191415] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:01.474 [2024-07-15 18:04:55.191455] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:01.474 [2024-07-15 18:04:55.191462] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:01.474 [2024-07-15 18:04:55.191469] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:01.474 [2024-07-15 18:04:55.191474] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:01.474 [2024-07-15 18:04:55.191492] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:19:02.431 18:04:55 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:19:02.431 18:04:55 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@862 -- # return 0 00:19:02.431 18:04:55 nvmf_tcp.nvmf_tls -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:19:02.431 18:04:55 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@728 -- # xtrace_disable 00:19:02.431 18:04:55 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:02.431 18:04:55 nvmf_tcp.nvmf_tls -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:02.431 18:04:55 nvmf_tcp.nvmf_tls -- target/tls.sh@177 -- # NOT setup_nvmf_tgt /tmp/tmp.BYHgc4Ja0k 00:19:02.431 18:04:55 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@648 -- # local es=0 00:19:02.431 18:04:55 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@650 -- # valid_exec_arg setup_nvmf_tgt /tmp/tmp.BYHgc4Ja0k 00:19:02.431 18:04:55 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@636 -- # local arg=setup_nvmf_tgt 00:19:02.431 18:04:55 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:19:02.431 18:04:55 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@640 -- # type -t setup_nvmf_tgt 00:19:02.431 18:04:55 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:19:02.431 18:04:55 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@651 -- # setup_nvmf_tgt /tmp/tmp.BYHgc4Ja0k 00:19:02.431 18:04:55 nvmf_tcp.nvmf_tls -- target/tls.sh@49 -- # local key=/tmp/tmp.BYHgc4Ja0k 00:19:02.431 18:04:55 nvmf_tcp.nvmf_tls -- target/tls.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:19:02.431 [2024-07-15 18:04:56.046599] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:02.431 18:04:56 nvmf_tcp.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:19:02.690 18:04:56 nvmf_tcp.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:19:02.690 [2024-07-15 18:04:56.391499] tcp.c: 966:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:02.690 [2024-07-15 18:04:56.391697] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:02.690 18:04:56 nvmf_tcp.nvmf_tls -- target/tls.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:19:02.948 malloc0 00:19:02.948 18:04:56 nvmf_tcp.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:19:03.207 18:04:56 nvmf_tcp.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.BYHgc4Ja0k 00:19:03.207 [2024-07-15 18:04:56.904991] tcp.c:3620:tcp_load_psk: *ERROR*: Incorrect permissions for PSK file 00:19:03.207 [2024-07-15 18:04:56.905019] tcp.c:3706:nvmf_tcp_subsystem_add_host: *ERROR*: Could not retrieve PSK from file 00:19:03.207 [2024-07-15 18:04:56.905061] subsystem.c:1051:spdk_nvmf_subsystem_add_host_ext: *ERROR*: Unable to add host to TCP transport 00:19:03.207 request: 00:19:03.207 { 00:19:03.207 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:03.207 "host": "nqn.2016-06.io.spdk:host1", 00:19:03.207 "psk": "/tmp/tmp.BYHgc4Ja0k", 00:19:03.207 "method": "nvmf_subsystem_add_host", 00:19:03.207 "req_id": 1 00:19:03.207 } 00:19:03.207 Got JSON-RPC error response 00:19:03.207 response: 00:19:03.207 { 00:19:03.207 "code": -32603, 00:19:03.207 "message": "Internal error" 00:19:03.207 } 00:19:03.207 18:04:56 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@651 -- # es=1 00:19:03.208 18:04:56 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:19:03.208 18:04:56 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:19:03.208 18:04:56 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:19:03.208 18:04:56 nvmf_tcp.nvmf_tls -- target/tls.sh@180 -- # killprocess 620283 00:19:03.208 18:04:56 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@948 -- # '[' -z 620283 ']' 00:19:03.208 18:04:56 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@952 -- # kill -0 620283 00:19:03.208 18:04:56 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # uname 00:19:03.208 18:04:56 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:19:03.208 18:04:56 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 620283 00:19:03.466 18:04:56 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:19:03.466 18:04:56 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:19:03.466 18:04:56 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@966 -- # echo 'killing process with pid 620283' 00:19:03.466 killing process with pid 620283 00:19:03.466 18:04:56 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@967 -- # kill 620283 00:19:03.466 18:04:56 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@972 -- # wait 620283 00:19:03.466 18:04:57 nvmf_tcp.nvmf_tls -- target/tls.sh@181 -- # chmod 0600 /tmp/tmp.BYHgc4Ja0k 00:19:03.466 18:04:57 nvmf_tcp.nvmf_tls -- target/tls.sh@184 -- # nvmfappstart -m 0x2 00:19:03.466 18:04:57 nvmf_tcp.nvmf_tls -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:19:03.466 18:04:57 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@722 -- # xtrace_disable 00:19:03.466 18:04:57 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:03.467 18:04:57 nvmf_tcp.nvmf_tls -- nvmf/common.sh@481 -- # nvmfpid=620641 00:19:03.467 18:04:57 nvmf_tcp.nvmf_tls -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:19:03.467 18:04:57 nvmf_tcp.nvmf_tls -- nvmf/common.sh@482 -- # waitforlisten 620641 00:19:03.467 18:04:57 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@829 -- # '[' -z 620641 ']' 00:19:03.467 18:04:57 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:03.467 18:04:57 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@834 -- # local max_retries=100 00:19:03.467 18:04:57 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:03.467 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:03.467 18:04:57 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@838 -- # xtrace_disable 00:19:03.467 18:04:57 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:03.725 [2024-07-15 18:04:57.215837] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:19:03.725 [2024-07-15 18:04:57.215882] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:03.725 EAL: No free 2048 kB hugepages reported on node 1 00:19:03.725 [2024-07-15 18:04:57.274090] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:03.725 [2024-07-15 18:04:57.351472] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:03.725 [2024-07-15 18:04:57.351512] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:03.725 [2024-07-15 18:04:57.351519] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:03.725 [2024-07-15 18:04:57.351525] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:03.725 [2024-07-15 18:04:57.351530] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:03.725 [2024-07-15 18:04:57.351553] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:19:04.659 18:04:58 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:19:04.659 18:04:58 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@862 -- # return 0 00:19:04.659 18:04:58 nvmf_tcp.nvmf_tls -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:19:04.659 18:04:58 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@728 -- # xtrace_disable 00:19:04.659 18:04:58 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:04.659 18:04:58 nvmf_tcp.nvmf_tls -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:04.659 18:04:58 nvmf_tcp.nvmf_tls -- target/tls.sh@185 -- # setup_nvmf_tgt /tmp/tmp.BYHgc4Ja0k 00:19:04.659 18:04:58 nvmf_tcp.nvmf_tls -- target/tls.sh@49 -- # local key=/tmp/tmp.BYHgc4Ja0k 00:19:04.659 18:04:58 nvmf_tcp.nvmf_tls -- target/tls.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:19:04.659 [2024-07-15 18:04:58.218571] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:04.659 18:04:58 nvmf_tcp.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:19:04.918 18:04:58 nvmf_tcp.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:19:04.918 [2024-07-15 18:04:58.547417] tcp.c: 966:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:04.918 [2024-07-15 18:04:58.547602] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:04.918 18:04:58 nvmf_tcp.nvmf_tls -- target/tls.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:19:05.178 malloc0 00:19:05.178 18:04:58 nvmf_tcp.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:19:05.437 18:04:58 nvmf_tcp.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.BYHgc4Ja0k 00:19:05.437 [2024-07-15 18:04:59.068981] tcp.c:3710:nvmf_tcp_subsystem_add_host: *WARNING*: nvmf_tcp_psk_path: deprecated feature PSK path to be removed in v24.09 00:19:05.437 18:04:59 nvmf_tcp.nvmf_tls -- target/tls.sh@187 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:19:05.437 18:04:59 nvmf_tcp.nvmf_tls -- target/tls.sh@188 -- # bdevperf_pid=621034 00:19:05.437 18:04:59 nvmf_tcp.nvmf_tls -- target/tls.sh@190 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:19:05.437 18:04:59 nvmf_tcp.nvmf_tls -- target/tls.sh@191 -- # waitforlisten 621034 /var/tmp/bdevperf.sock 00:19:05.437 18:04:59 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@829 -- # '[' -z 621034 ']' 00:19:05.437 18:04:59 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:05.437 18:04:59 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@834 -- # local max_retries=100 00:19:05.437 18:04:59 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:05.437 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:05.437 18:04:59 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@838 -- # xtrace_disable 00:19:05.437 18:04:59 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:05.437 [2024-07-15 18:04:59.119005] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:19:05.437 [2024-07-15 18:04:59.119051] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid621034 ] 00:19:05.437 EAL: No free 2048 kB hugepages reported on node 1 00:19:05.696 [2024-07-15 18:04:59.168851] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:05.696 [2024-07-15 18:04:59.242497] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:19:05.696 18:04:59 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:19:05.696 18:04:59 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@862 -- # return 0 00:19:05.696 18:04:59 nvmf_tcp.nvmf_tls -- target/tls.sh@192 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.BYHgc4Ja0k 00:19:05.953 [2024-07-15 18:04:59.482453] bdev_nvme_rpc.c: 517:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:05.953 [2024-07-15 18:04:59.482526] nvme_tcp.c:2589:nvme_tcp_generate_tls_credentials: *WARNING*: nvme_ctrlr_psk: deprecated feature spdk_nvme_ctrlr_opts.psk to be removed in v24.09 00:19:05.953 TLSTESTn1 00:19:05.953 18:04:59 nvmf_tcp.nvmf_tls -- target/tls.sh@196 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py save_config 00:19:06.210 18:04:59 nvmf_tcp.nvmf_tls -- target/tls.sh@196 -- # tgtconf='{ 00:19:06.210 "subsystems": [ 00:19:06.210 { 00:19:06.210 "subsystem": "keyring", 00:19:06.210 "config": [] 00:19:06.210 }, 00:19:06.210 { 00:19:06.210 "subsystem": "iobuf", 00:19:06.210 "config": [ 00:19:06.210 { 00:19:06.210 "method": "iobuf_set_options", 00:19:06.210 "params": { 00:19:06.210 "small_pool_count": 8192, 00:19:06.210 "large_pool_count": 1024, 00:19:06.210 "small_bufsize": 8192, 00:19:06.210 "large_bufsize": 135168 00:19:06.210 } 00:19:06.210 } 00:19:06.210 ] 00:19:06.210 }, 00:19:06.210 { 00:19:06.210 "subsystem": "sock", 00:19:06.210 "config": [ 00:19:06.210 { 00:19:06.210 "method": "sock_set_default_impl", 00:19:06.210 "params": { 00:19:06.210 "impl_name": "posix" 00:19:06.210 } 00:19:06.210 }, 00:19:06.210 { 00:19:06.210 "method": "sock_impl_set_options", 00:19:06.210 "params": { 00:19:06.210 "impl_name": "ssl", 00:19:06.210 "recv_buf_size": 4096, 00:19:06.210 "send_buf_size": 4096, 00:19:06.210 "enable_recv_pipe": true, 00:19:06.210 "enable_quickack": false, 00:19:06.210 "enable_placement_id": 0, 00:19:06.210 "enable_zerocopy_send_server": true, 00:19:06.210 "enable_zerocopy_send_client": false, 00:19:06.210 "zerocopy_threshold": 0, 00:19:06.210 "tls_version": 0, 00:19:06.210 "enable_ktls": false 00:19:06.210 } 00:19:06.210 }, 00:19:06.210 { 00:19:06.210 "method": "sock_impl_set_options", 00:19:06.210 "params": { 00:19:06.211 "impl_name": "posix", 00:19:06.211 "recv_buf_size": 2097152, 00:19:06.211 "send_buf_size": 2097152, 00:19:06.211 "enable_recv_pipe": true, 00:19:06.211 "enable_quickack": false, 00:19:06.211 "enable_placement_id": 0, 00:19:06.211 "enable_zerocopy_send_server": true, 00:19:06.211 "enable_zerocopy_send_client": false, 00:19:06.211 "zerocopy_threshold": 0, 00:19:06.211 "tls_version": 0, 00:19:06.211 "enable_ktls": false 00:19:06.211 } 00:19:06.211 } 00:19:06.211 ] 00:19:06.211 }, 00:19:06.211 { 00:19:06.211 "subsystem": "vmd", 00:19:06.211 "config": [] 00:19:06.211 }, 00:19:06.211 { 00:19:06.211 "subsystem": "accel", 00:19:06.211 "config": [ 00:19:06.211 { 00:19:06.211 "method": "accel_set_options", 00:19:06.211 "params": { 00:19:06.211 "small_cache_size": 128, 00:19:06.211 "large_cache_size": 16, 00:19:06.211 "task_count": 2048, 00:19:06.211 "sequence_count": 2048, 00:19:06.211 "buf_count": 2048 00:19:06.211 } 00:19:06.211 } 00:19:06.211 ] 00:19:06.211 }, 00:19:06.211 { 00:19:06.211 "subsystem": "bdev", 00:19:06.211 "config": [ 00:19:06.211 { 00:19:06.211 "method": "bdev_set_options", 00:19:06.211 "params": { 00:19:06.211 "bdev_io_pool_size": 65535, 00:19:06.211 "bdev_io_cache_size": 256, 00:19:06.211 "bdev_auto_examine": true, 00:19:06.211 "iobuf_small_cache_size": 128, 00:19:06.211 "iobuf_large_cache_size": 16 00:19:06.211 } 00:19:06.211 }, 00:19:06.211 { 00:19:06.211 "method": "bdev_raid_set_options", 00:19:06.211 "params": { 00:19:06.211 "process_window_size_kb": 1024 00:19:06.211 } 00:19:06.211 }, 00:19:06.211 { 00:19:06.211 "method": "bdev_iscsi_set_options", 00:19:06.211 "params": { 00:19:06.211 "timeout_sec": 30 00:19:06.211 } 00:19:06.211 }, 00:19:06.211 { 00:19:06.211 "method": "bdev_nvme_set_options", 00:19:06.211 "params": { 00:19:06.211 "action_on_timeout": "none", 00:19:06.211 "timeout_us": 0, 00:19:06.211 "timeout_admin_us": 0, 00:19:06.211 "keep_alive_timeout_ms": 10000, 00:19:06.211 "arbitration_burst": 0, 00:19:06.211 "low_priority_weight": 0, 00:19:06.211 "medium_priority_weight": 0, 00:19:06.211 "high_priority_weight": 0, 00:19:06.211 "nvme_adminq_poll_period_us": 10000, 00:19:06.211 "nvme_ioq_poll_period_us": 0, 00:19:06.211 "io_queue_requests": 0, 00:19:06.211 "delay_cmd_submit": true, 00:19:06.211 "transport_retry_count": 4, 00:19:06.211 "bdev_retry_count": 3, 00:19:06.211 "transport_ack_timeout": 0, 00:19:06.211 "ctrlr_loss_timeout_sec": 0, 00:19:06.211 "reconnect_delay_sec": 0, 00:19:06.211 "fast_io_fail_timeout_sec": 0, 00:19:06.211 "disable_auto_failback": false, 00:19:06.211 "generate_uuids": false, 00:19:06.211 "transport_tos": 0, 00:19:06.211 "nvme_error_stat": false, 00:19:06.211 "rdma_srq_size": 0, 00:19:06.211 "io_path_stat": false, 00:19:06.211 "allow_accel_sequence": false, 00:19:06.211 "rdma_max_cq_size": 0, 00:19:06.211 "rdma_cm_event_timeout_ms": 0, 00:19:06.211 "dhchap_digests": [ 00:19:06.211 "sha256", 00:19:06.211 "sha384", 00:19:06.211 "sha512" 00:19:06.211 ], 00:19:06.211 "dhchap_dhgroups": [ 00:19:06.211 "null", 00:19:06.211 "ffdhe2048", 00:19:06.211 "ffdhe3072", 00:19:06.211 "ffdhe4096", 00:19:06.211 "ffdhe6144", 00:19:06.211 "ffdhe8192" 00:19:06.211 ] 00:19:06.211 } 00:19:06.211 }, 00:19:06.211 { 00:19:06.211 "method": "bdev_nvme_set_hotplug", 00:19:06.211 "params": { 00:19:06.211 "period_us": 100000, 00:19:06.211 "enable": false 00:19:06.211 } 00:19:06.211 }, 00:19:06.211 { 00:19:06.211 "method": "bdev_malloc_create", 00:19:06.211 "params": { 00:19:06.211 "name": "malloc0", 00:19:06.211 "num_blocks": 8192, 00:19:06.211 "block_size": 4096, 00:19:06.211 "physical_block_size": 4096, 00:19:06.211 "uuid": "83ab41b9-db37-4fb1-a8ae-c075b6c5463e", 00:19:06.211 "optimal_io_boundary": 0 00:19:06.211 } 00:19:06.211 }, 00:19:06.211 { 00:19:06.211 "method": "bdev_wait_for_examine" 00:19:06.211 } 00:19:06.211 ] 00:19:06.211 }, 00:19:06.211 { 00:19:06.211 "subsystem": "nbd", 00:19:06.211 "config": [] 00:19:06.211 }, 00:19:06.211 { 00:19:06.211 "subsystem": "scheduler", 00:19:06.211 "config": [ 00:19:06.211 { 00:19:06.211 "method": "framework_set_scheduler", 00:19:06.211 "params": { 00:19:06.211 "name": "static" 00:19:06.211 } 00:19:06.211 } 00:19:06.211 ] 00:19:06.211 }, 00:19:06.211 { 00:19:06.211 "subsystem": "nvmf", 00:19:06.211 "config": [ 00:19:06.211 { 00:19:06.211 "method": "nvmf_set_config", 00:19:06.211 "params": { 00:19:06.211 "discovery_filter": "match_any", 00:19:06.211 "admin_cmd_passthru": { 00:19:06.211 "identify_ctrlr": false 00:19:06.211 } 00:19:06.211 } 00:19:06.211 }, 00:19:06.211 { 00:19:06.211 "method": "nvmf_set_max_subsystems", 00:19:06.211 "params": { 00:19:06.211 "max_subsystems": 1024 00:19:06.211 } 00:19:06.211 }, 00:19:06.211 { 00:19:06.211 "method": "nvmf_set_crdt", 00:19:06.211 "params": { 00:19:06.211 "crdt1": 0, 00:19:06.211 "crdt2": 0, 00:19:06.211 "crdt3": 0 00:19:06.211 } 00:19:06.211 }, 00:19:06.211 { 00:19:06.211 "method": "nvmf_create_transport", 00:19:06.211 "params": { 00:19:06.211 "trtype": "TCP", 00:19:06.211 "max_queue_depth": 128, 00:19:06.211 "max_io_qpairs_per_ctrlr": 127, 00:19:06.211 "in_capsule_data_size": 4096, 00:19:06.211 "max_io_size": 131072, 00:19:06.211 "io_unit_size": 131072, 00:19:06.211 "max_aq_depth": 128, 00:19:06.211 "num_shared_buffers": 511, 00:19:06.211 "buf_cache_size": 4294967295, 00:19:06.211 "dif_insert_or_strip": false, 00:19:06.211 "zcopy": false, 00:19:06.211 "c2h_success": false, 00:19:06.211 "sock_priority": 0, 00:19:06.211 "abort_timeout_sec": 1, 00:19:06.211 "ack_timeout": 0, 00:19:06.211 "data_wr_pool_size": 0 00:19:06.211 } 00:19:06.211 }, 00:19:06.211 { 00:19:06.211 "method": "nvmf_create_subsystem", 00:19:06.211 "params": { 00:19:06.211 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:06.211 "allow_any_host": false, 00:19:06.211 "serial_number": "SPDK00000000000001", 00:19:06.211 "model_number": "SPDK bdev Controller", 00:19:06.211 "max_namespaces": 10, 00:19:06.211 "min_cntlid": 1, 00:19:06.211 "max_cntlid": 65519, 00:19:06.211 "ana_reporting": false 00:19:06.211 } 00:19:06.211 }, 00:19:06.211 { 00:19:06.211 "method": "nvmf_subsystem_add_host", 00:19:06.211 "params": { 00:19:06.211 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:06.211 "host": "nqn.2016-06.io.spdk:host1", 00:19:06.211 "psk": "/tmp/tmp.BYHgc4Ja0k" 00:19:06.211 } 00:19:06.211 }, 00:19:06.211 { 00:19:06.211 "method": "nvmf_subsystem_add_ns", 00:19:06.211 "params": { 00:19:06.211 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:06.211 "namespace": { 00:19:06.211 "nsid": 1, 00:19:06.211 "bdev_name": "malloc0", 00:19:06.211 "nguid": "83AB41B9DB374FB1A8AEC075B6C5463E", 00:19:06.211 "uuid": "83ab41b9-db37-4fb1-a8ae-c075b6c5463e", 00:19:06.211 "no_auto_visible": false 00:19:06.211 } 00:19:06.211 } 00:19:06.211 }, 00:19:06.211 { 00:19:06.211 "method": "nvmf_subsystem_add_listener", 00:19:06.211 "params": { 00:19:06.211 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:06.211 "listen_address": { 00:19:06.211 "trtype": "TCP", 00:19:06.211 "adrfam": "IPv4", 00:19:06.211 "traddr": "10.0.0.2", 00:19:06.211 "trsvcid": "4420" 00:19:06.211 }, 00:19:06.211 "secure_channel": true 00:19:06.211 } 00:19:06.211 } 00:19:06.211 ] 00:19:06.211 } 00:19:06.211 ] 00:19:06.211 }' 00:19:06.211 18:04:59 nvmf_tcp.nvmf_tls -- target/tls.sh@197 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock save_config 00:19:06.469 18:05:00 nvmf_tcp.nvmf_tls -- target/tls.sh@197 -- # bdevperfconf='{ 00:19:06.469 "subsystems": [ 00:19:06.469 { 00:19:06.469 "subsystem": "keyring", 00:19:06.469 "config": [] 00:19:06.469 }, 00:19:06.469 { 00:19:06.469 "subsystem": "iobuf", 00:19:06.469 "config": [ 00:19:06.469 { 00:19:06.469 "method": "iobuf_set_options", 00:19:06.469 "params": { 00:19:06.469 "small_pool_count": 8192, 00:19:06.469 "large_pool_count": 1024, 00:19:06.469 "small_bufsize": 8192, 00:19:06.469 "large_bufsize": 135168 00:19:06.469 } 00:19:06.469 } 00:19:06.469 ] 00:19:06.469 }, 00:19:06.469 { 00:19:06.469 "subsystem": "sock", 00:19:06.469 "config": [ 00:19:06.469 { 00:19:06.469 "method": "sock_set_default_impl", 00:19:06.469 "params": { 00:19:06.469 "impl_name": "posix" 00:19:06.469 } 00:19:06.469 }, 00:19:06.469 { 00:19:06.469 "method": "sock_impl_set_options", 00:19:06.469 "params": { 00:19:06.469 "impl_name": "ssl", 00:19:06.469 "recv_buf_size": 4096, 00:19:06.469 "send_buf_size": 4096, 00:19:06.469 "enable_recv_pipe": true, 00:19:06.469 "enable_quickack": false, 00:19:06.469 "enable_placement_id": 0, 00:19:06.469 "enable_zerocopy_send_server": true, 00:19:06.469 "enable_zerocopy_send_client": false, 00:19:06.469 "zerocopy_threshold": 0, 00:19:06.469 "tls_version": 0, 00:19:06.469 "enable_ktls": false 00:19:06.469 } 00:19:06.469 }, 00:19:06.469 { 00:19:06.469 "method": "sock_impl_set_options", 00:19:06.469 "params": { 00:19:06.469 "impl_name": "posix", 00:19:06.469 "recv_buf_size": 2097152, 00:19:06.469 "send_buf_size": 2097152, 00:19:06.469 "enable_recv_pipe": true, 00:19:06.469 "enable_quickack": false, 00:19:06.469 "enable_placement_id": 0, 00:19:06.469 "enable_zerocopy_send_server": true, 00:19:06.469 "enable_zerocopy_send_client": false, 00:19:06.469 "zerocopy_threshold": 0, 00:19:06.469 "tls_version": 0, 00:19:06.469 "enable_ktls": false 00:19:06.469 } 00:19:06.469 } 00:19:06.469 ] 00:19:06.469 }, 00:19:06.469 { 00:19:06.469 "subsystem": "vmd", 00:19:06.469 "config": [] 00:19:06.469 }, 00:19:06.469 { 00:19:06.469 "subsystem": "accel", 00:19:06.469 "config": [ 00:19:06.469 { 00:19:06.469 "method": "accel_set_options", 00:19:06.469 "params": { 00:19:06.469 "small_cache_size": 128, 00:19:06.469 "large_cache_size": 16, 00:19:06.469 "task_count": 2048, 00:19:06.469 "sequence_count": 2048, 00:19:06.469 "buf_count": 2048 00:19:06.469 } 00:19:06.469 } 00:19:06.469 ] 00:19:06.469 }, 00:19:06.469 { 00:19:06.469 "subsystem": "bdev", 00:19:06.469 "config": [ 00:19:06.469 { 00:19:06.469 "method": "bdev_set_options", 00:19:06.469 "params": { 00:19:06.469 "bdev_io_pool_size": 65535, 00:19:06.469 "bdev_io_cache_size": 256, 00:19:06.469 "bdev_auto_examine": true, 00:19:06.469 "iobuf_small_cache_size": 128, 00:19:06.469 "iobuf_large_cache_size": 16 00:19:06.469 } 00:19:06.469 }, 00:19:06.469 { 00:19:06.469 "method": "bdev_raid_set_options", 00:19:06.469 "params": { 00:19:06.469 "process_window_size_kb": 1024 00:19:06.469 } 00:19:06.469 }, 00:19:06.469 { 00:19:06.469 "method": "bdev_iscsi_set_options", 00:19:06.469 "params": { 00:19:06.469 "timeout_sec": 30 00:19:06.469 } 00:19:06.469 }, 00:19:06.469 { 00:19:06.469 "method": "bdev_nvme_set_options", 00:19:06.469 "params": { 00:19:06.469 "action_on_timeout": "none", 00:19:06.469 "timeout_us": 0, 00:19:06.469 "timeout_admin_us": 0, 00:19:06.469 "keep_alive_timeout_ms": 10000, 00:19:06.469 "arbitration_burst": 0, 00:19:06.469 "low_priority_weight": 0, 00:19:06.469 "medium_priority_weight": 0, 00:19:06.469 "high_priority_weight": 0, 00:19:06.469 "nvme_adminq_poll_period_us": 10000, 00:19:06.469 "nvme_ioq_poll_period_us": 0, 00:19:06.469 "io_queue_requests": 512, 00:19:06.469 "delay_cmd_submit": true, 00:19:06.469 "transport_retry_count": 4, 00:19:06.469 "bdev_retry_count": 3, 00:19:06.469 "transport_ack_timeout": 0, 00:19:06.469 "ctrlr_loss_timeout_sec": 0, 00:19:06.469 "reconnect_delay_sec": 0, 00:19:06.469 "fast_io_fail_timeout_sec": 0, 00:19:06.469 "disable_auto_failback": false, 00:19:06.469 "generate_uuids": false, 00:19:06.469 "transport_tos": 0, 00:19:06.469 "nvme_error_stat": false, 00:19:06.469 "rdma_srq_size": 0, 00:19:06.469 "io_path_stat": false, 00:19:06.469 "allow_accel_sequence": false, 00:19:06.469 "rdma_max_cq_size": 0, 00:19:06.469 "rdma_cm_event_timeout_ms": 0, 00:19:06.469 "dhchap_digests": [ 00:19:06.469 "sha256", 00:19:06.469 "sha384", 00:19:06.469 "sha512" 00:19:06.469 ], 00:19:06.469 "dhchap_dhgroups": [ 00:19:06.469 "null", 00:19:06.469 "ffdhe2048", 00:19:06.469 "ffdhe3072", 00:19:06.469 "ffdhe4096", 00:19:06.469 "ffdhe6144", 00:19:06.469 "ffdhe8192" 00:19:06.469 ] 00:19:06.469 } 00:19:06.469 }, 00:19:06.469 { 00:19:06.469 "method": "bdev_nvme_attach_controller", 00:19:06.469 "params": { 00:19:06.469 "name": "TLSTEST", 00:19:06.470 "trtype": "TCP", 00:19:06.470 "adrfam": "IPv4", 00:19:06.470 "traddr": "10.0.0.2", 00:19:06.470 "trsvcid": "4420", 00:19:06.470 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:19:06.470 "prchk_reftag": false, 00:19:06.470 "prchk_guard": false, 00:19:06.470 "ctrlr_loss_timeout_sec": 0, 00:19:06.470 "reconnect_delay_sec": 0, 00:19:06.470 "fast_io_fail_timeout_sec": 0, 00:19:06.470 "psk": "/tmp/tmp.BYHgc4Ja0k", 00:19:06.470 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:19:06.470 "hdgst": false, 00:19:06.470 "ddgst": false 00:19:06.470 } 00:19:06.470 }, 00:19:06.470 { 00:19:06.470 "method": "bdev_nvme_set_hotplug", 00:19:06.470 "params": { 00:19:06.470 "period_us": 100000, 00:19:06.470 "enable": false 00:19:06.470 } 00:19:06.470 }, 00:19:06.470 { 00:19:06.470 "method": "bdev_wait_for_examine" 00:19:06.470 } 00:19:06.470 ] 00:19:06.470 }, 00:19:06.470 { 00:19:06.470 "subsystem": "nbd", 00:19:06.470 "config": [] 00:19:06.470 } 00:19:06.470 ] 00:19:06.470 }' 00:19:06.470 18:05:00 nvmf_tcp.nvmf_tls -- target/tls.sh@199 -- # killprocess 621034 00:19:06.470 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@948 -- # '[' -z 621034 ']' 00:19:06.470 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@952 -- # kill -0 621034 00:19:06.470 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # uname 00:19:06.470 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:19:06.470 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 621034 00:19:06.470 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # process_name=reactor_2 00:19:06.470 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@958 -- # '[' reactor_2 = sudo ']' 00:19:06.470 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@966 -- # echo 'killing process with pid 621034' 00:19:06.470 killing process with pid 621034 00:19:06.470 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@967 -- # kill 621034 00:19:06.470 Received shutdown signal, test time was about 10.000000 seconds 00:19:06.470 00:19:06.470 Latency(us) 00:19:06.470 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:06.470 =================================================================================================================== 00:19:06.470 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:19:06.470 [2024-07-15 18:05:00.103501] app.c:1024:log_deprecation_hits: *WARNING*: nvme_ctrlr_psk: deprecation 'spdk_nvme_ctrlr_opts.psk' scheduled for removal in v24.09 hit 1 times 00:19:06.470 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@972 -- # wait 621034 00:19:06.728 18:05:00 nvmf_tcp.nvmf_tls -- target/tls.sh@200 -- # killprocess 620641 00:19:06.728 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@948 -- # '[' -z 620641 ']' 00:19:06.728 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@952 -- # kill -0 620641 00:19:06.728 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # uname 00:19:06.728 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:19:06.728 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 620641 00:19:06.728 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:19:06.728 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:19:06.728 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@966 -- # echo 'killing process with pid 620641' 00:19:06.728 killing process with pid 620641 00:19:06.728 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@967 -- # kill 620641 00:19:06.728 [2024-07-15 18:05:00.330774] app.c:1024:log_deprecation_hits: *WARNING*: nvmf_tcp_psk_path: deprecation 'PSK path' scheduled for removal in v24.09 hit 1 times 00:19:06.728 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@972 -- # wait 620641 00:19:06.987 18:05:00 nvmf_tcp.nvmf_tls -- target/tls.sh@203 -- # nvmfappstart -m 0x2 -c /dev/fd/62 00:19:06.987 18:05:00 nvmf_tcp.nvmf_tls -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:19:06.987 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@722 -- # xtrace_disable 00:19:06.987 18:05:00 nvmf_tcp.nvmf_tls -- target/tls.sh@203 -- # echo '{ 00:19:06.987 "subsystems": [ 00:19:06.987 { 00:19:06.987 "subsystem": "keyring", 00:19:06.987 "config": [] 00:19:06.987 }, 00:19:06.987 { 00:19:06.987 "subsystem": "iobuf", 00:19:06.987 "config": [ 00:19:06.987 { 00:19:06.987 "method": "iobuf_set_options", 00:19:06.987 "params": { 00:19:06.987 "small_pool_count": 8192, 00:19:06.987 "large_pool_count": 1024, 00:19:06.987 "small_bufsize": 8192, 00:19:06.987 "large_bufsize": 135168 00:19:06.987 } 00:19:06.987 } 00:19:06.987 ] 00:19:06.987 }, 00:19:06.987 { 00:19:06.987 "subsystem": "sock", 00:19:06.987 "config": [ 00:19:06.987 { 00:19:06.987 "method": "sock_set_default_impl", 00:19:06.987 "params": { 00:19:06.987 "impl_name": "posix" 00:19:06.987 } 00:19:06.987 }, 00:19:06.987 { 00:19:06.987 "method": "sock_impl_set_options", 00:19:06.987 "params": { 00:19:06.987 "impl_name": "ssl", 00:19:06.987 "recv_buf_size": 4096, 00:19:06.987 "send_buf_size": 4096, 00:19:06.987 "enable_recv_pipe": true, 00:19:06.987 "enable_quickack": false, 00:19:06.987 "enable_placement_id": 0, 00:19:06.987 "enable_zerocopy_send_server": true, 00:19:06.987 "enable_zerocopy_send_client": false, 00:19:06.987 "zerocopy_threshold": 0, 00:19:06.987 "tls_version": 0, 00:19:06.987 "enable_ktls": false 00:19:06.987 } 00:19:06.987 }, 00:19:06.987 { 00:19:06.987 "method": "sock_impl_set_options", 00:19:06.987 "params": { 00:19:06.988 "impl_name": "posix", 00:19:06.988 "recv_buf_size": 2097152, 00:19:06.988 "send_buf_size": 2097152, 00:19:06.988 "enable_recv_pipe": true, 00:19:06.988 "enable_quickack": false, 00:19:06.988 "enable_placement_id": 0, 00:19:06.988 "enable_zerocopy_send_server": true, 00:19:06.988 "enable_zerocopy_send_client": false, 00:19:06.988 "zerocopy_threshold": 0, 00:19:06.988 "tls_version": 0, 00:19:06.988 "enable_ktls": false 00:19:06.988 } 00:19:06.988 } 00:19:06.988 ] 00:19:06.988 }, 00:19:06.988 { 00:19:06.988 "subsystem": "vmd", 00:19:06.988 "config": [] 00:19:06.988 }, 00:19:06.988 { 00:19:06.988 "subsystem": "accel", 00:19:06.988 "config": [ 00:19:06.988 { 00:19:06.988 "method": "accel_set_options", 00:19:06.988 "params": { 00:19:06.988 "small_cache_size": 128, 00:19:06.988 "large_cache_size": 16, 00:19:06.988 "task_count": 2048, 00:19:06.988 "sequence_count": 2048, 00:19:06.988 "buf_count": 2048 00:19:06.988 } 00:19:06.988 } 00:19:06.988 ] 00:19:06.988 }, 00:19:06.988 { 00:19:06.988 "subsystem": "bdev", 00:19:06.988 "config": [ 00:19:06.988 { 00:19:06.988 "method": "bdev_set_options", 00:19:06.988 "params": { 00:19:06.988 "bdev_io_pool_size": 65535, 00:19:06.988 "bdev_io_cache_size": 256, 00:19:06.988 "bdev_auto_examine": true, 00:19:06.988 "iobuf_small_cache_size": 128, 00:19:06.988 "iobuf_large_cache_size": 16 00:19:06.988 } 00:19:06.988 }, 00:19:06.988 { 00:19:06.988 "method": "bdev_raid_set_options", 00:19:06.988 "params": { 00:19:06.988 "process_window_size_kb": 1024 00:19:06.988 } 00:19:06.988 }, 00:19:06.988 { 00:19:06.988 "method": "bdev_iscsi_set_options", 00:19:06.988 "params": { 00:19:06.988 "timeout_sec": 30 00:19:06.988 } 00:19:06.988 }, 00:19:06.988 { 00:19:06.988 "method": "bdev_nvme_set_options", 00:19:06.988 "params": { 00:19:06.988 "action_on_timeout": "none", 00:19:06.988 "timeout_us": 0, 00:19:06.988 "timeout_admin_us": 0, 00:19:06.988 "keep_alive_timeout_ms": 10000, 00:19:06.988 "arbitration_burst": 0, 00:19:06.988 "low_priority_weight": 0, 00:19:06.988 "medium_priority_weight": 0, 00:19:06.988 "high_priority_weight": 0, 00:19:06.988 "nvme_adminq_poll_period_us": 10000, 00:19:06.988 "nvme_ioq_poll_period_us": 0, 00:19:06.988 "io_queue_requests": 0, 00:19:06.988 "delay_cmd_submit": true, 00:19:06.988 "transport_retry_count": 4, 00:19:06.988 "bdev_retry_count": 3, 00:19:06.988 "transport_ack_timeout": 0, 00:19:06.988 "ctrlr_loss_timeout_sec": 0, 00:19:06.988 "reconnect_delay_sec": 0, 00:19:06.988 "fast_io_fail_timeout_sec": 0, 00:19:06.988 "disable_auto_failback": false, 00:19:06.988 "generate_uuids": false, 00:19:06.988 "transport_tos": 0, 00:19:06.988 "nvme_error_stat": false, 00:19:06.988 "rdma_srq_size": 0, 00:19:06.988 "io_path_stat": false, 00:19:06.988 "allow_accel_sequence": false, 00:19:06.988 "rdma_max_cq_size": 0, 00:19:06.988 "rdma_cm_event_timeout_ms": 0, 00:19:06.988 "dhchap_digests": [ 00:19:06.988 "sha256", 00:19:06.988 "sha384", 00:19:06.988 "sha512" 00:19:06.988 ], 00:19:06.988 "dhchap_dhgroups": [ 00:19:06.988 "null", 00:19:06.988 "ffdhe2048", 00:19:06.988 "ffdhe3072", 00:19:06.988 "ffdhe4096", 00:19:06.988 "ffdhe6144", 00:19:06.988 "ffdhe8192" 00:19:06.988 ] 00:19:06.988 } 00:19:06.988 }, 00:19:06.988 { 00:19:06.988 "method": "bdev_nvme_set_hotplug", 00:19:06.988 "params": { 00:19:06.988 "period_us": 100000, 00:19:06.988 "enable": false 00:19:06.988 } 00:19:06.988 }, 00:19:06.988 { 00:19:06.988 "method": "bdev_malloc_create", 00:19:06.988 "params": { 00:19:06.988 "name": "malloc0", 00:19:06.988 "num_blocks": 8192, 00:19:06.988 "block_size": 4096, 00:19:06.988 "physical_block_size": 4096, 00:19:06.988 "uuid": "83ab41b9-db37-4fb1-a8ae-c075b6c5463e", 00:19:06.988 "optimal_io_boundary": 0 00:19:06.988 } 00:19:06.988 }, 00:19:06.988 { 00:19:06.988 "method": "bdev_wait_for_examine" 00:19:06.988 } 00:19:06.988 ] 00:19:06.988 }, 00:19:06.988 { 00:19:06.988 "subsystem": "nbd", 00:19:06.988 "config": [] 00:19:06.988 }, 00:19:06.988 { 00:19:06.988 "subsystem": "scheduler", 00:19:06.988 "config": [ 00:19:06.988 { 00:19:06.988 "method": "framework_set_scheduler", 00:19:06.988 "params": { 00:19:06.988 "name": "static" 00:19:06.988 } 00:19:06.988 } 00:19:06.988 ] 00:19:06.988 }, 00:19:06.988 { 00:19:06.988 "subsystem": "nvmf", 00:19:06.988 "config": [ 00:19:06.988 { 00:19:06.988 "method": "nvmf_set_config", 00:19:06.988 "params": { 00:19:06.988 "discovery_filter": "match_any", 00:19:06.988 "admin_cmd_passthru": { 00:19:06.988 "identify_ctrlr": false 00:19:06.988 } 00:19:06.988 } 00:19:06.988 }, 00:19:06.988 { 00:19:06.988 "method": "nvmf_set_max_subsystems", 00:19:06.988 "params": { 00:19:06.988 "max_subsystems": 1024 00:19:06.988 } 00:19:06.988 }, 00:19:06.988 { 00:19:06.988 "method": "nvmf_set_crdt", 00:19:06.988 "params": { 00:19:06.988 "crdt1": 0, 00:19:06.988 "crdt2": 0, 00:19:06.988 "crdt3": 0 00:19:06.988 } 00:19:06.988 }, 00:19:06.988 { 00:19:06.988 "method": "nvmf_create_transport", 00:19:06.988 "params": { 00:19:06.988 "trtype": "TCP", 00:19:06.988 "max_queue_depth": 128, 00:19:06.988 "max_io_qpairs_per_ctrlr": 127, 00:19:06.988 "in_capsule_data_size": 4096, 00:19:06.988 "max_io_size": 131072, 00:19:06.988 "io_unit_size": 131072, 00:19:06.988 "max_aq_depth": 128, 00:19:06.988 "num_shared_buffers": 511, 00:19:06.988 "buf_cache_size": 4294967295, 00:19:06.988 "dif_insert_or_strip": false, 00:19:06.988 "zcopy": false, 00:19:06.988 "c2h_success": false, 00:19:06.988 "sock_priority": 0, 00:19:06.988 "abort_timeout_sec": 1, 00:19:06.988 "ack_timeout": 0, 00:19:06.988 "data_wr_pool_size": 0 00:19:06.988 } 00:19:06.988 }, 00:19:06.988 { 00:19:06.988 "method": "nvmf_create_subsystem", 00:19:06.988 "params": { 00:19:06.988 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:06.988 "allow_any_host": false, 00:19:06.988 "serial_number": "SPDK00000000000001", 00:19:06.988 "model_number": "SPDK bdev Controller", 00:19:06.988 "max_namespaces": 10, 00:19:06.988 "min_cntlid": 1, 00:19:06.988 "max_cntlid": 65519, 00:19:06.988 "ana_reporting": false 00:19:06.988 } 00:19:06.988 }, 00:19:06.988 { 00:19:06.988 "method": "nvmf_subsystem_add_host", 00:19:06.988 "params": { 00:19:06.988 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:06.988 "host": "nqn.2016-06.io.spdk:host1", 00:19:06.988 "psk": "/tmp/tmp.BYHgc4Ja0k" 00:19:06.988 } 00:19:06.988 }, 00:19:06.988 { 00:19:06.988 "method": "nvmf_subsystem_add_ns", 00:19:06.988 "params": { 00:19:06.988 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:06.988 "namespace": { 00:19:06.988 "nsid": 1, 00:19:06.988 "bdev_name": "malloc0", 00:19:06.988 "nguid": "83AB41B9DB374FB1A8AEC075B6C5463E", 00:19:06.988 "uuid": "83ab41b9-db37-4fb1-a8ae-c075b6c5463e", 00:19:06.988 "no_auto_visible": false 00:19:06.988 } 00:19:06.988 } 00:19:06.988 }, 00:19:06.988 { 00:19:06.988 "method": "nvmf_subsystem_add_listener", 00:19:06.988 "params": { 00:19:06.988 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:06.988 "listen_address": { 00:19:06.988 "trtype": "TCP", 00:19:06.988 "adrfam": "IPv4", 00:19:06.988 "traddr": "10.0.0.2", 00:19:06.988 "trsvcid": "4420" 00:19:06.988 }, 00:19:06.988 "secure_channel": true 00:19:06.988 } 00:19:06.988 } 00:19:06.988 ] 00:19:06.988 } 00:19:06.988 ] 00:19:06.988 }' 00:19:06.988 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:06.988 18:05:00 nvmf_tcp.nvmf_tls -- nvmf/common.sh@481 -- # nvmfpid=621277 00:19:06.988 18:05:00 nvmf_tcp.nvmf_tls -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 -c /dev/fd/62 00:19:06.988 18:05:00 nvmf_tcp.nvmf_tls -- nvmf/common.sh@482 -- # waitforlisten 621277 00:19:06.988 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@829 -- # '[' -z 621277 ']' 00:19:06.988 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:06.988 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@834 -- # local max_retries=100 00:19:06.989 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:06.989 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:06.989 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@838 -- # xtrace_disable 00:19:06.989 18:05:00 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:06.989 [2024-07-15 18:05:00.579507] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:19:06.989 [2024-07-15 18:05:00.579551] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:06.989 EAL: No free 2048 kB hugepages reported on node 1 00:19:06.989 [2024-07-15 18:05:00.637934] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:07.248 [2024-07-15 18:05:00.716137] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:07.248 [2024-07-15 18:05:00.716174] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:07.248 [2024-07-15 18:05:00.716182] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:07.248 [2024-07-15 18:05:00.716187] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:07.248 [2024-07-15 18:05:00.716193] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:07.248 [2024-07-15 18:05:00.716258] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:19:07.248 [2024-07-15 18:05:00.917312] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:07.248 [2024-07-15 18:05:00.941916] tcp.c:3710:nvmf_tcp_subsystem_add_host: *WARNING*: nvmf_tcp_psk_path: deprecated feature PSK path to be removed in v24.09 00:19:07.248 [2024-07-15 18:05:00.957964] tcp.c: 966:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:07.248 [2024-07-15 18:05:00.958136] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:07.815 18:05:01 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:19:07.815 18:05:01 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@862 -- # return 0 00:19:07.815 18:05:01 nvmf_tcp.nvmf_tls -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:19:07.815 18:05:01 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@728 -- # xtrace_disable 00:19:07.815 18:05:01 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:07.815 18:05:01 nvmf_tcp.nvmf_tls -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:07.815 18:05:01 nvmf_tcp.nvmf_tls -- target/tls.sh@207 -- # bdevperf_pid=621433 00:19:07.815 18:05:01 nvmf_tcp.nvmf_tls -- target/tls.sh@208 -- # waitforlisten 621433 /var/tmp/bdevperf.sock 00:19:07.815 18:05:01 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@829 -- # '[' -z 621433 ']' 00:19:07.815 18:05:01 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:07.815 18:05:01 nvmf_tcp.nvmf_tls -- target/tls.sh@204 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 -c /dev/fd/63 00:19:07.815 18:05:01 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@834 -- # local max_retries=100 00:19:07.815 18:05:01 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:07.815 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:07.815 18:05:01 nvmf_tcp.nvmf_tls -- target/tls.sh@204 -- # echo '{ 00:19:07.815 "subsystems": [ 00:19:07.815 { 00:19:07.815 "subsystem": "keyring", 00:19:07.815 "config": [] 00:19:07.815 }, 00:19:07.815 { 00:19:07.815 "subsystem": "iobuf", 00:19:07.815 "config": [ 00:19:07.815 { 00:19:07.815 "method": "iobuf_set_options", 00:19:07.815 "params": { 00:19:07.815 "small_pool_count": 8192, 00:19:07.815 "large_pool_count": 1024, 00:19:07.815 "small_bufsize": 8192, 00:19:07.815 "large_bufsize": 135168 00:19:07.815 } 00:19:07.815 } 00:19:07.815 ] 00:19:07.815 }, 00:19:07.815 { 00:19:07.815 "subsystem": "sock", 00:19:07.815 "config": [ 00:19:07.815 { 00:19:07.815 "method": "sock_set_default_impl", 00:19:07.815 "params": { 00:19:07.815 "impl_name": "posix" 00:19:07.815 } 00:19:07.815 }, 00:19:07.815 { 00:19:07.815 "method": "sock_impl_set_options", 00:19:07.815 "params": { 00:19:07.815 "impl_name": "ssl", 00:19:07.815 "recv_buf_size": 4096, 00:19:07.815 "send_buf_size": 4096, 00:19:07.815 "enable_recv_pipe": true, 00:19:07.815 "enable_quickack": false, 00:19:07.815 "enable_placement_id": 0, 00:19:07.815 "enable_zerocopy_send_server": true, 00:19:07.815 "enable_zerocopy_send_client": false, 00:19:07.815 "zerocopy_threshold": 0, 00:19:07.815 "tls_version": 0, 00:19:07.815 "enable_ktls": false 00:19:07.815 } 00:19:07.815 }, 00:19:07.815 { 00:19:07.815 "method": "sock_impl_set_options", 00:19:07.815 "params": { 00:19:07.815 "impl_name": "posix", 00:19:07.815 "recv_buf_size": 2097152, 00:19:07.815 "send_buf_size": 2097152, 00:19:07.815 "enable_recv_pipe": true, 00:19:07.815 "enable_quickack": false, 00:19:07.815 "enable_placement_id": 0, 00:19:07.815 "enable_zerocopy_send_server": true, 00:19:07.815 "enable_zerocopy_send_client": false, 00:19:07.815 "zerocopy_threshold": 0, 00:19:07.815 "tls_version": 0, 00:19:07.815 "enable_ktls": false 00:19:07.815 } 00:19:07.815 } 00:19:07.815 ] 00:19:07.815 }, 00:19:07.815 { 00:19:07.815 "subsystem": "vmd", 00:19:07.815 "config": [] 00:19:07.815 }, 00:19:07.816 { 00:19:07.816 "subsystem": "accel", 00:19:07.816 "config": [ 00:19:07.816 { 00:19:07.816 "method": "accel_set_options", 00:19:07.816 "params": { 00:19:07.816 "small_cache_size": 128, 00:19:07.816 "large_cache_size": 16, 00:19:07.816 "task_count": 2048, 00:19:07.816 "sequence_count": 2048, 00:19:07.816 "buf_count": 2048 00:19:07.816 } 00:19:07.816 } 00:19:07.816 ] 00:19:07.816 }, 00:19:07.816 { 00:19:07.816 "subsystem": "bdev", 00:19:07.816 "config": [ 00:19:07.816 { 00:19:07.816 "method": "bdev_set_options", 00:19:07.816 "params": { 00:19:07.816 "bdev_io_pool_size": 65535, 00:19:07.816 "bdev_io_cache_size": 256, 00:19:07.816 "bdev_auto_examine": true, 00:19:07.816 "iobuf_small_cache_size": 128, 00:19:07.816 "iobuf_large_cache_size": 16 00:19:07.816 } 00:19:07.816 }, 00:19:07.816 { 00:19:07.816 "method": "bdev_raid_set_options", 00:19:07.816 "params": { 00:19:07.816 "process_window_size_kb": 1024 00:19:07.816 } 00:19:07.816 }, 00:19:07.816 { 00:19:07.816 "method": "bdev_iscsi_set_options", 00:19:07.816 "params": { 00:19:07.816 "timeout_sec": 30 00:19:07.816 } 00:19:07.816 }, 00:19:07.816 { 00:19:07.816 "method": "bdev_nvme_set_options", 00:19:07.816 "params": { 00:19:07.816 "action_on_timeout": "none", 00:19:07.816 "timeout_us": 0, 00:19:07.816 "timeout_admin_us": 0, 00:19:07.816 "keep_alive_timeout_ms": 10000, 00:19:07.816 "arbitration_burst": 0, 00:19:07.816 "low_priority_weight": 0, 00:19:07.816 "medium_priority_weight": 0, 00:19:07.816 "high_priority_weight": 0, 00:19:07.816 "nvme_adminq_poll_period_us": 10000, 00:19:07.816 "nvme_ioq_poll_period_us": 0, 00:19:07.816 "io_queue_requests": 512, 00:19:07.816 "delay_cmd_submit": true, 00:19:07.816 "transport_retry_count": 4, 00:19:07.816 "bdev_retry_count": 3, 00:19:07.816 "transport_ack_timeout": 0, 00:19:07.816 "ctrlr_loss_timeout_sec": 0, 00:19:07.816 "reconnect_delay_sec": 0, 00:19:07.816 "fast_io_fail_timeout_sec": 0, 00:19:07.816 "disable_auto_failback": false, 00:19:07.816 "generate_uuids": false, 00:19:07.816 "transport_tos": 0, 00:19:07.816 "nvme_error_stat": false, 00:19:07.816 "rdma_srq_size": 0, 00:19:07.816 "io_path_stat": false, 00:19:07.816 "allow_accel_sequence": false, 00:19:07.816 "rdma_max_cq_size": 0, 00:19:07.816 "rdma_cm_event_timeout_ms": 0, 00:19:07.816 "dhchap_digests": [ 00:19:07.816 "sha256", 00:19:07.816 "sha384", 00:19:07.816 "sha512" 00:19:07.816 ], 00:19:07.816 "dhchap_dhgroups": [ 00:19:07.816 "null", 00:19:07.816 "ffdhe2048", 00:19:07.816 "ffdhe3072", 00:19:07.816 "ffdhe4096", 00:19:07.816 "ffdhe6144", 00:19:07.816 "ffdhe8192" 00:19:07.816 ] 00:19:07.816 } 00:19:07.816 }, 00:19:07.816 { 00:19:07.816 "method": "bdev_nvme_attach_controller", 00:19:07.816 "params": { 00:19:07.816 "name": "TLSTEST", 00:19:07.816 "trtype": "TCP", 00:19:07.816 "adrfam": "IPv4", 00:19:07.816 "traddr": "10.0.0.2", 00:19:07.816 "trsvcid": "4420", 00:19:07.816 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:19:07.816 "prchk_reftag": false, 00:19:07.816 "prchk_guard": false, 00:19:07.816 "ctrlr_loss_timeout_sec": 0, 00:19:07.816 "reconnect_delay_sec": 0, 00:19:07.816 "fast_io_fail_timeout_sec": 0, 00:19:07.816 "psk": "/tmp/tmp.BYHgc4Ja0k", 00:19:07.816 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:19:07.816 "hdgst": false, 00:19:07.816 "ddgst": false 00:19:07.816 } 00:19:07.816 }, 00:19:07.816 { 00:19:07.816 "method": "bdev_nvme_set_hotplug", 00:19:07.816 "params": { 00:19:07.816 "period_us": 100000, 00:19:07.816 "enable": false 00:19:07.816 } 00:19:07.816 }, 00:19:07.816 { 00:19:07.816 "method": "bdev_wait_for_examine" 00:19:07.816 } 00:19:07.816 ] 00:19:07.816 }, 00:19:07.816 { 00:19:07.816 "subsystem": "nbd", 00:19:07.816 "config": [] 00:19:07.816 } 00:19:07.816 ] 00:19:07.816 }' 00:19:07.816 18:05:01 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@838 -- # xtrace_disable 00:19:07.816 18:05:01 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:07.816 [2024-07-15 18:05:01.461587] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:19:07.816 [2024-07-15 18:05:01.461637] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid621433 ] 00:19:07.816 EAL: No free 2048 kB hugepages reported on node 1 00:19:07.816 [2024-07-15 18:05:01.512711] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:08.075 [2024-07-15 18:05:01.589850] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:19:08.075 [2024-07-15 18:05:01.732500] bdev_nvme_rpc.c: 517:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:08.075 [2024-07-15 18:05:01.732578] nvme_tcp.c:2589:nvme_tcp_generate_tls_credentials: *WARNING*: nvme_ctrlr_psk: deprecated feature spdk_nvme_ctrlr_opts.psk to be removed in v24.09 00:19:08.642 18:05:02 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:19:08.642 18:05:02 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@862 -- # return 0 00:19:08.642 18:05:02 nvmf_tcp.nvmf_tls -- target/tls.sh@211 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 20 -s /var/tmp/bdevperf.sock perform_tests 00:19:08.642 Running I/O for 10 seconds... 00:19:20.863 00:19:20.863 Latency(us) 00:19:20.863 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:20.863 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:19:20.863 Verification LBA range: start 0x0 length 0x2000 00:19:20.863 TLSTESTn1 : 10.02 4507.64 17.61 0.00 0.00 28345.86 6012.22 54480.36 00:19:20.863 =================================================================================================================== 00:19:20.863 Total : 4507.64 17.61 0.00 0.00 28345.86 6012.22 54480.36 00:19:20.863 0 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- target/tls.sh@213 -- # trap 'nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- target/tls.sh@214 -- # killprocess 621433 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@948 -- # '[' -z 621433 ']' 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@952 -- # kill -0 621433 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # uname 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 621433 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # process_name=reactor_2 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@958 -- # '[' reactor_2 = sudo ']' 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@966 -- # echo 'killing process with pid 621433' 00:19:20.863 killing process with pid 621433 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@967 -- # kill 621433 00:19:20.863 Received shutdown signal, test time was about 10.000000 seconds 00:19:20.863 00:19:20.863 Latency(us) 00:19:20.863 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:20.863 =================================================================================================================== 00:19:20.863 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:19:20.863 [2024-07-15 18:05:12.454961] app.c:1024:log_deprecation_hits: *WARNING*: nvme_ctrlr_psk: deprecation 'spdk_nvme_ctrlr_opts.psk' scheduled for removal in v24.09 hit 1 times 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@972 -- # wait 621433 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- target/tls.sh@215 -- # killprocess 621277 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@948 -- # '[' -z 621277 ']' 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@952 -- # kill -0 621277 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # uname 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 621277 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@966 -- # echo 'killing process with pid 621277' 00:19:20.863 killing process with pid 621277 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@967 -- # kill 621277 00:19:20.863 [2024-07-15 18:05:12.678478] app.c:1024:log_deprecation_hits: *WARNING*: nvmf_tcp_psk_path: deprecation 'PSK path' scheduled for removal in v24.09 hit 1 times 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@972 -- # wait 621277 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- target/tls.sh@218 -- # nvmfappstart 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@722 -- # xtrace_disable 00:19:20.863 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:20.864 18:05:12 nvmf_tcp.nvmf_tls -- nvmf/common.sh@481 -- # nvmfpid=623362 00:19:20.864 18:05:12 nvmf_tcp.nvmf_tls -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:19:20.864 18:05:12 nvmf_tcp.nvmf_tls -- nvmf/common.sh@482 -- # waitforlisten 623362 00:19:20.864 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@829 -- # '[' -z 623362 ']' 00:19:20.864 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:20.864 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@834 -- # local max_retries=100 00:19:20.864 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:20.864 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:20.864 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@838 -- # xtrace_disable 00:19:20.864 18:05:12 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:20.864 [2024-07-15 18:05:12.924505] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:19:20.864 [2024-07-15 18:05:12.924550] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:20.864 EAL: No free 2048 kB hugepages reported on node 1 00:19:20.864 [2024-07-15 18:05:12.981779] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:20.864 [2024-07-15 18:05:13.049673] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:20.864 [2024-07-15 18:05:13.049713] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:20.864 [2024-07-15 18:05:13.049720] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:20.864 [2024-07-15 18:05:13.049729] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:20.864 [2024-07-15 18:05:13.049734] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:20.864 [2024-07-15 18:05:13.049774] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:19:20.864 18:05:13 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:19:20.864 18:05:13 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@862 -- # return 0 00:19:20.864 18:05:13 nvmf_tcp.nvmf_tls -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:19:20.864 18:05:13 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@728 -- # xtrace_disable 00:19:20.864 18:05:13 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:20.864 18:05:13 nvmf_tcp.nvmf_tls -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:20.864 18:05:13 nvmf_tcp.nvmf_tls -- target/tls.sh@219 -- # setup_nvmf_tgt /tmp/tmp.BYHgc4Ja0k 00:19:20.864 18:05:13 nvmf_tcp.nvmf_tls -- target/tls.sh@49 -- # local key=/tmp/tmp.BYHgc4Ja0k 00:19:20.864 18:05:13 nvmf_tcp.nvmf_tls -- target/tls.sh@51 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:19:20.864 [2024-07-15 18:05:13.917192] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:20.864 18:05:13 nvmf_tcp.nvmf_tls -- target/tls.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -s SPDK00000000000001 -m 10 00:19:20.864 18:05:14 nvmf_tcp.nvmf_tls -- target/tls.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -k 00:19:20.864 [2024-07-15 18:05:14.270085] tcp.c: 966:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:20.864 [2024-07-15 18:05:14.270300] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:20.864 18:05:14 nvmf_tcp.nvmf_tls -- target/tls.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 32 4096 -b malloc0 00:19:20.864 malloc0 00:19:20.864 18:05:14 nvmf_tcp.nvmf_tls -- target/tls.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 malloc0 -n 1 00:19:21.123 18:05:14 nvmf_tcp.nvmf_tls -- target/tls.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode1 nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.BYHgc4Ja0k 00:19:21.123 [2024-07-15 18:05:14.779517] tcp.c:3710:nvmf_tcp_subsystem_add_host: *WARNING*: nvmf_tcp_psk_path: deprecated feature PSK path to be removed in v24.09 00:19:21.123 18:05:14 nvmf_tcp.nvmf_tls -- target/tls.sh@220 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -z -r /var/tmp/bdevperf.sock -q 128 -o 4k -w verify -t 1 00:19:21.123 18:05:14 nvmf_tcp.nvmf_tls -- target/tls.sh@222 -- # bdevperf_pid=623626 00:19:21.123 18:05:14 nvmf_tcp.nvmf_tls -- target/tls.sh@224 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:19:21.123 18:05:14 nvmf_tcp.nvmf_tls -- target/tls.sh@225 -- # waitforlisten 623626 /var/tmp/bdevperf.sock 00:19:21.123 18:05:14 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@829 -- # '[' -z 623626 ']' 00:19:21.123 18:05:14 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:21.123 18:05:14 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@834 -- # local max_retries=100 00:19:21.123 18:05:14 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:21.123 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:21.123 18:05:14 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@838 -- # xtrace_disable 00:19:21.123 18:05:14 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:21.123 [2024-07-15 18:05:14.830567] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:19:21.123 [2024-07-15 18:05:14.830615] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid623626 ] 00:19:21.382 EAL: No free 2048 kB hugepages reported on node 1 00:19:21.382 [2024-07-15 18:05:14.884095] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:21.382 [2024-07-15 18:05:14.958683] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:19:21.949 18:05:15 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:19:21.949 18:05:15 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@862 -- # return 0 00:19:21.949 18:05:15 nvmf_tcp.nvmf_tls -- target/tls.sh@227 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.BYHgc4Ja0k 00:19:22.208 18:05:15 nvmf_tcp.nvmf_tls -- target/tls.sh@228 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 --psk key0 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 00:19:22.500 [2024-07-15 18:05:15.962081] bdev_nvme_rpc.c: 517:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:22.500 nvme0n1 00:19:22.500 18:05:16 nvmf_tcp.nvmf_tls -- target/tls.sh@232 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:19:22.500 Running I/O for 1 seconds... 00:19:23.472 00:19:23.472 Latency(us) 00:19:23.472 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:23.472 Job: nvme0n1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:19:23.472 Verification LBA range: start 0x0 length 0x2000 00:19:23.472 nvme0n1 : 1.01 5068.44 19.80 0.00 0.00 25055.70 4729.99 31001.38 00:19:23.472 =================================================================================================================== 00:19:23.472 Total : 5068.44 19.80 0.00 0.00 25055.70 4729.99 31001.38 00:19:23.472 0 00:19:23.472 18:05:17 nvmf_tcp.nvmf_tls -- target/tls.sh@234 -- # killprocess 623626 00:19:23.472 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@948 -- # '[' -z 623626 ']' 00:19:23.472 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@952 -- # kill -0 623626 00:19:23.472 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # uname 00:19:23.472 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:19:23.472 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 623626 00:19:23.472 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:19:23.472 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:19:23.472 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@966 -- # echo 'killing process with pid 623626' 00:19:23.472 killing process with pid 623626 00:19:23.472 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@967 -- # kill 623626 00:19:23.472 Received shutdown signal, test time was about 1.000000 seconds 00:19:23.472 00:19:23.472 Latency(us) 00:19:23.472 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:23.472 =================================================================================================================== 00:19:23.472 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:19:23.472 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@972 -- # wait 623626 00:19:23.731 18:05:17 nvmf_tcp.nvmf_tls -- target/tls.sh@235 -- # killprocess 623362 00:19:23.731 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@948 -- # '[' -z 623362 ']' 00:19:23.731 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@952 -- # kill -0 623362 00:19:23.731 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # uname 00:19:23.731 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:19:23.731 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 623362 00:19:23.731 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:19:23.731 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:19:23.731 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@966 -- # echo 'killing process with pid 623362' 00:19:23.731 killing process with pid 623362 00:19:23.731 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@967 -- # kill 623362 00:19:23.731 [2024-07-15 18:05:17.419878] app.c:1024:log_deprecation_hits: *WARNING*: nvmf_tcp_psk_path: deprecation 'PSK path' scheduled for removal in v24.09 hit 1 times 00:19:23.731 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@972 -- # wait 623362 00:19:23.990 18:05:17 nvmf_tcp.nvmf_tls -- target/tls.sh@238 -- # nvmfappstart 00:19:23.990 18:05:17 nvmf_tcp.nvmf_tls -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:19:23.990 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@722 -- # xtrace_disable 00:19:23.990 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:23.990 18:05:17 nvmf_tcp.nvmf_tls -- nvmf/common.sh@481 -- # nvmfpid=624102 00:19:23.990 18:05:17 nvmf_tcp.nvmf_tls -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:19:23.990 18:05:17 nvmf_tcp.nvmf_tls -- nvmf/common.sh@482 -- # waitforlisten 624102 00:19:23.990 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@829 -- # '[' -z 624102 ']' 00:19:23.990 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:23.990 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@834 -- # local max_retries=100 00:19:23.990 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:23.990 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:23.990 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@838 -- # xtrace_disable 00:19:23.990 18:05:17 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:23.990 [2024-07-15 18:05:17.666050] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:19:23.990 [2024-07-15 18:05:17.666095] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:23.990 EAL: No free 2048 kB hugepages reported on node 1 00:19:24.249 [2024-07-15 18:05:17.722807] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:24.249 [2024-07-15 18:05:17.801209] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:24.249 [2024-07-15 18:05:17.801252] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:24.249 [2024-07-15 18:05:17.801259] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:24.249 [2024-07-15 18:05:17.801265] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:24.249 [2024-07-15 18:05:17.801269] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:24.249 [2024-07-15 18:05:17.801310] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:19:24.816 18:05:18 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:19:24.816 18:05:18 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@862 -- # return 0 00:19:24.816 18:05:18 nvmf_tcp.nvmf_tls -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:19:24.816 18:05:18 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@728 -- # xtrace_disable 00:19:24.816 18:05:18 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:24.816 18:05:18 nvmf_tcp.nvmf_tls -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:24.816 18:05:18 nvmf_tcp.nvmf_tls -- target/tls.sh@239 -- # rpc_cmd 00:19:24.816 18:05:18 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@559 -- # xtrace_disable 00:19:24.816 18:05:18 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:24.816 [2024-07-15 18:05:18.515463] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:24.816 malloc0 00:19:24.816 [2024-07-15 18:05:18.543745] tcp.c: 966:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:24.816 [2024-07-15 18:05:18.543951] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:25.076 18:05:18 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:19:25.076 18:05:18 nvmf_tcp.nvmf_tls -- target/tls.sh@252 -- # bdevperf_pid=624344 00:19:25.076 18:05:18 nvmf_tcp.nvmf_tls -- target/tls.sh@254 -- # waitforlisten 624344 /var/tmp/bdevperf.sock 00:19:25.076 18:05:18 nvmf_tcp.nvmf_tls -- target/tls.sh@250 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -z -r /var/tmp/bdevperf.sock -q 128 -o 4k -w verify -t 1 00:19:25.076 18:05:18 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@829 -- # '[' -z 624344 ']' 00:19:25.076 18:05:18 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:25.076 18:05:18 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@834 -- # local max_retries=100 00:19:25.076 18:05:18 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:25.076 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:25.076 18:05:18 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@838 -- # xtrace_disable 00:19:25.076 18:05:18 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:25.076 [2024-07-15 18:05:18.617554] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:19:25.076 [2024-07-15 18:05:18.617594] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid624344 ] 00:19:25.076 EAL: No free 2048 kB hugepages reported on node 1 00:19:25.076 [2024-07-15 18:05:18.671165] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:25.076 [2024-07-15 18:05:18.745828] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:19:26.012 18:05:19 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:19:26.012 18:05:19 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@862 -- # return 0 00:19:26.012 18:05:19 nvmf_tcp.nvmf_tls -- target/tls.sh@255 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock keyring_file_add_key key0 /tmp/tmp.BYHgc4Ja0k 00:19:26.012 18:05:19 nvmf_tcp.nvmf_tls -- target/tls.sh@256 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 --psk key0 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 00:19:26.270 [2024-07-15 18:05:19.745847] bdev_nvme_rpc.c: 517:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:26.270 nvme0n1 00:19:26.270 18:05:19 nvmf_tcp.nvmf_tls -- target/tls.sh@260 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:19:26.270 Running I/O for 1 seconds... 00:19:27.647 00:19:27.647 Latency(us) 00:19:27.647 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:27.647 Job: nvme0n1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:19:27.647 Verification LBA range: start 0x0 length 0x2000 00:19:27.647 nvme0n1 : 1.01 5005.18 19.55 0.00 0.00 25370.27 6126.19 56531.92 00:19:27.647 =================================================================================================================== 00:19:27.647 Total : 5005.18 19.55 0.00 0.00 25370.27 6126.19 56531.92 00:19:27.647 0 00:19:27.647 18:05:20 nvmf_tcp.nvmf_tls -- target/tls.sh@263 -- # rpc_cmd save_config 00:19:27.647 18:05:20 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@559 -- # xtrace_disable 00:19:27.647 18:05:20 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:27.647 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:19:27.647 18:05:21 nvmf_tcp.nvmf_tls -- target/tls.sh@263 -- # tgtcfg='{ 00:19:27.647 "subsystems": [ 00:19:27.647 { 00:19:27.647 "subsystem": "keyring", 00:19:27.647 "config": [ 00:19:27.647 { 00:19:27.647 "method": "keyring_file_add_key", 00:19:27.647 "params": { 00:19:27.647 "name": "key0", 00:19:27.647 "path": "/tmp/tmp.BYHgc4Ja0k" 00:19:27.647 } 00:19:27.647 } 00:19:27.647 ] 00:19:27.647 }, 00:19:27.647 { 00:19:27.647 "subsystem": "iobuf", 00:19:27.647 "config": [ 00:19:27.647 { 00:19:27.647 "method": "iobuf_set_options", 00:19:27.647 "params": { 00:19:27.647 "small_pool_count": 8192, 00:19:27.647 "large_pool_count": 1024, 00:19:27.647 "small_bufsize": 8192, 00:19:27.647 "large_bufsize": 135168 00:19:27.647 } 00:19:27.647 } 00:19:27.647 ] 00:19:27.647 }, 00:19:27.647 { 00:19:27.647 "subsystem": "sock", 00:19:27.647 "config": [ 00:19:27.647 { 00:19:27.647 "method": "sock_set_default_impl", 00:19:27.647 "params": { 00:19:27.647 "impl_name": "posix" 00:19:27.647 } 00:19:27.647 }, 00:19:27.647 { 00:19:27.647 "method": "sock_impl_set_options", 00:19:27.647 "params": { 00:19:27.647 "impl_name": "ssl", 00:19:27.647 "recv_buf_size": 4096, 00:19:27.647 "send_buf_size": 4096, 00:19:27.647 "enable_recv_pipe": true, 00:19:27.647 "enable_quickack": false, 00:19:27.647 "enable_placement_id": 0, 00:19:27.647 "enable_zerocopy_send_server": true, 00:19:27.647 "enable_zerocopy_send_client": false, 00:19:27.647 "zerocopy_threshold": 0, 00:19:27.647 "tls_version": 0, 00:19:27.647 "enable_ktls": false 00:19:27.647 } 00:19:27.647 }, 00:19:27.647 { 00:19:27.647 "method": "sock_impl_set_options", 00:19:27.647 "params": { 00:19:27.647 "impl_name": "posix", 00:19:27.647 "recv_buf_size": 2097152, 00:19:27.647 "send_buf_size": 2097152, 00:19:27.647 "enable_recv_pipe": true, 00:19:27.647 "enable_quickack": false, 00:19:27.647 "enable_placement_id": 0, 00:19:27.647 "enable_zerocopy_send_server": true, 00:19:27.647 "enable_zerocopy_send_client": false, 00:19:27.647 "zerocopy_threshold": 0, 00:19:27.647 "tls_version": 0, 00:19:27.647 "enable_ktls": false 00:19:27.647 } 00:19:27.647 } 00:19:27.647 ] 00:19:27.647 }, 00:19:27.647 { 00:19:27.647 "subsystem": "vmd", 00:19:27.647 "config": [] 00:19:27.647 }, 00:19:27.647 { 00:19:27.647 "subsystem": "accel", 00:19:27.647 "config": [ 00:19:27.647 { 00:19:27.647 "method": "accel_set_options", 00:19:27.647 "params": { 00:19:27.647 "small_cache_size": 128, 00:19:27.647 "large_cache_size": 16, 00:19:27.647 "task_count": 2048, 00:19:27.647 "sequence_count": 2048, 00:19:27.647 "buf_count": 2048 00:19:27.647 } 00:19:27.647 } 00:19:27.647 ] 00:19:27.647 }, 00:19:27.647 { 00:19:27.647 "subsystem": "bdev", 00:19:27.647 "config": [ 00:19:27.647 { 00:19:27.647 "method": "bdev_set_options", 00:19:27.647 "params": { 00:19:27.647 "bdev_io_pool_size": 65535, 00:19:27.647 "bdev_io_cache_size": 256, 00:19:27.647 "bdev_auto_examine": true, 00:19:27.647 "iobuf_small_cache_size": 128, 00:19:27.647 "iobuf_large_cache_size": 16 00:19:27.647 } 00:19:27.647 }, 00:19:27.647 { 00:19:27.647 "method": "bdev_raid_set_options", 00:19:27.647 "params": { 00:19:27.647 "process_window_size_kb": 1024 00:19:27.647 } 00:19:27.647 }, 00:19:27.647 { 00:19:27.647 "method": "bdev_iscsi_set_options", 00:19:27.647 "params": { 00:19:27.647 "timeout_sec": 30 00:19:27.647 } 00:19:27.647 }, 00:19:27.647 { 00:19:27.647 "method": "bdev_nvme_set_options", 00:19:27.647 "params": { 00:19:27.647 "action_on_timeout": "none", 00:19:27.647 "timeout_us": 0, 00:19:27.647 "timeout_admin_us": 0, 00:19:27.647 "keep_alive_timeout_ms": 10000, 00:19:27.647 "arbitration_burst": 0, 00:19:27.647 "low_priority_weight": 0, 00:19:27.647 "medium_priority_weight": 0, 00:19:27.647 "high_priority_weight": 0, 00:19:27.647 "nvme_adminq_poll_period_us": 10000, 00:19:27.648 "nvme_ioq_poll_period_us": 0, 00:19:27.648 "io_queue_requests": 0, 00:19:27.648 "delay_cmd_submit": true, 00:19:27.648 "transport_retry_count": 4, 00:19:27.648 "bdev_retry_count": 3, 00:19:27.648 "transport_ack_timeout": 0, 00:19:27.648 "ctrlr_loss_timeout_sec": 0, 00:19:27.648 "reconnect_delay_sec": 0, 00:19:27.648 "fast_io_fail_timeout_sec": 0, 00:19:27.648 "disable_auto_failback": false, 00:19:27.648 "generate_uuids": false, 00:19:27.648 "transport_tos": 0, 00:19:27.648 "nvme_error_stat": false, 00:19:27.648 "rdma_srq_size": 0, 00:19:27.648 "io_path_stat": false, 00:19:27.648 "allow_accel_sequence": false, 00:19:27.648 "rdma_max_cq_size": 0, 00:19:27.648 "rdma_cm_event_timeout_ms": 0, 00:19:27.648 "dhchap_digests": [ 00:19:27.648 "sha256", 00:19:27.648 "sha384", 00:19:27.648 "sha512" 00:19:27.648 ], 00:19:27.648 "dhchap_dhgroups": [ 00:19:27.648 "null", 00:19:27.648 "ffdhe2048", 00:19:27.648 "ffdhe3072", 00:19:27.648 "ffdhe4096", 00:19:27.648 "ffdhe6144", 00:19:27.648 "ffdhe8192" 00:19:27.648 ] 00:19:27.648 } 00:19:27.648 }, 00:19:27.648 { 00:19:27.648 "method": "bdev_nvme_set_hotplug", 00:19:27.648 "params": { 00:19:27.648 "period_us": 100000, 00:19:27.648 "enable": false 00:19:27.648 } 00:19:27.648 }, 00:19:27.648 { 00:19:27.648 "method": "bdev_malloc_create", 00:19:27.648 "params": { 00:19:27.648 "name": "malloc0", 00:19:27.648 "num_blocks": 8192, 00:19:27.648 "block_size": 4096, 00:19:27.648 "physical_block_size": 4096, 00:19:27.648 "uuid": "5038965e-4a63-4e67-a174-38221306e7a2", 00:19:27.648 "optimal_io_boundary": 0 00:19:27.648 } 00:19:27.648 }, 00:19:27.648 { 00:19:27.648 "method": "bdev_wait_for_examine" 00:19:27.648 } 00:19:27.648 ] 00:19:27.648 }, 00:19:27.648 { 00:19:27.648 "subsystem": "nbd", 00:19:27.648 "config": [] 00:19:27.648 }, 00:19:27.648 { 00:19:27.648 "subsystem": "scheduler", 00:19:27.648 "config": [ 00:19:27.648 { 00:19:27.648 "method": "framework_set_scheduler", 00:19:27.648 "params": { 00:19:27.648 "name": "static" 00:19:27.648 } 00:19:27.648 } 00:19:27.648 ] 00:19:27.648 }, 00:19:27.648 { 00:19:27.648 "subsystem": "nvmf", 00:19:27.648 "config": [ 00:19:27.648 { 00:19:27.648 "method": "nvmf_set_config", 00:19:27.648 "params": { 00:19:27.648 "discovery_filter": "match_any", 00:19:27.648 "admin_cmd_passthru": { 00:19:27.648 "identify_ctrlr": false 00:19:27.648 } 00:19:27.648 } 00:19:27.648 }, 00:19:27.648 { 00:19:27.648 "method": "nvmf_set_max_subsystems", 00:19:27.648 "params": { 00:19:27.648 "max_subsystems": 1024 00:19:27.648 } 00:19:27.648 }, 00:19:27.648 { 00:19:27.648 "method": "nvmf_set_crdt", 00:19:27.648 "params": { 00:19:27.648 "crdt1": 0, 00:19:27.648 "crdt2": 0, 00:19:27.648 "crdt3": 0 00:19:27.648 } 00:19:27.648 }, 00:19:27.648 { 00:19:27.648 "method": "nvmf_create_transport", 00:19:27.648 "params": { 00:19:27.648 "trtype": "TCP", 00:19:27.648 "max_queue_depth": 128, 00:19:27.648 "max_io_qpairs_per_ctrlr": 127, 00:19:27.648 "in_capsule_data_size": 4096, 00:19:27.648 "max_io_size": 131072, 00:19:27.648 "io_unit_size": 131072, 00:19:27.648 "max_aq_depth": 128, 00:19:27.648 "num_shared_buffers": 511, 00:19:27.648 "buf_cache_size": 4294967295, 00:19:27.648 "dif_insert_or_strip": false, 00:19:27.648 "zcopy": false, 00:19:27.648 "c2h_success": false, 00:19:27.648 "sock_priority": 0, 00:19:27.648 "abort_timeout_sec": 1, 00:19:27.648 "ack_timeout": 0, 00:19:27.648 "data_wr_pool_size": 0 00:19:27.648 } 00:19:27.648 }, 00:19:27.648 { 00:19:27.648 "method": "nvmf_create_subsystem", 00:19:27.648 "params": { 00:19:27.648 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:27.648 "allow_any_host": false, 00:19:27.648 "serial_number": "00000000000000000000", 00:19:27.648 "model_number": "SPDK bdev Controller", 00:19:27.648 "max_namespaces": 32, 00:19:27.648 "min_cntlid": 1, 00:19:27.648 "max_cntlid": 65519, 00:19:27.648 "ana_reporting": false 00:19:27.648 } 00:19:27.648 }, 00:19:27.648 { 00:19:27.648 "method": "nvmf_subsystem_add_host", 00:19:27.648 "params": { 00:19:27.648 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:27.648 "host": "nqn.2016-06.io.spdk:host1", 00:19:27.648 "psk": "key0" 00:19:27.648 } 00:19:27.648 }, 00:19:27.648 { 00:19:27.648 "method": "nvmf_subsystem_add_ns", 00:19:27.648 "params": { 00:19:27.648 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:27.648 "namespace": { 00:19:27.648 "nsid": 1, 00:19:27.648 "bdev_name": "malloc0", 00:19:27.648 "nguid": "5038965E4A634E67A17438221306E7A2", 00:19:27.648 "uuid": "5038965e-4a63-4e67-a174-38221306e7a2", 00:19:27.648 "no_auto_visible": false 00:19:27.648 } 00:19:27.648 } 00:19:27.648 }, 00:19:27.648 { 00:19:27.648 "method": "nvmf_subsystem_add_listener", 00:19:27.648 "params": { 00:19:27.648 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:27.648 "listen_address": { 00:19:27.648 "trtype": "TCP", 00:19:27.648 "adrfam": "IPv4", 00:19:27.648 "traddr": "10.0.0.2", 00:19:27.648 "trsvcid": "4420" 00:19:27.648 }, 00:19:27.648 "secure_channel": true 00:19:27.648 } 00:19:27.648 } 00:19:27.648 ] 00:19:27.648 } 00:19:27.648 ] 00:19:27.648 }' 00:19:27.648 18:05:21 nvmf_tcp.nvmf_tls -- target/tls.sh@264 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock save_config 00:19:27.648 18:05:21 nvmf_tcp.nvmf_tls -- target/tls.sh@264 -- # bperfcfg='{ 00:19:27.648 "subsystems": [ 00:19:27.648 { 00:19:27.648 "subsystem": "keyring", 00:19:27.648 "config": [ 00:19:27.648 { 00:19:27.648 "method": "keyring_file_add_key", 00:19:27.648 "params": { 00:19:27.648 "name": "key0", 00:19:27.648 "path": "/tmp/tmp.BYHgc4Ja0k" 00:19:27.648 } 00:19:27.648 } 00:19:27.648 ] 00:19:27.648 }, 00:19:27.648 { 00:19:27.648 "subsystem": "iobuf", 00:19:27.648 "config": [ 00:19:27.648 { 00:19:27.648 "method": "iobuf_set_options", 00:19:27.648 "params": { 00:19:27.648 "small_pool_count": 8192, 00:19:27.648 "large_pool_count": 1024, 00:19:27.648 "small_bufsize": 8192, 00:19:27.648 "large_bufsize": 135168 00:19:27.648 } 00:19:27.648 } 00:19:27.648 ] 00:19:27.648 }, 00:19:27.648 { 00:19:27.648 "subsystem": "sock", 00:19:27.648 "config": [ 00:19:27.648 { 00:19:27.648 "method": "sock_set_default_impl", 00:19:27.648 "params": { 00:19:27.648 "impl_name": "posix" 00:19:27.648 } 00:19:27.648 }, 00:19:27.648 { 00:19:27.648 "method": "sock_impl_set_options", 00:19:27.648 "params": { 00:19:27.648 "impl_name": "ssl", 00:19:27.648 "recv_buf_size": 4096, 00:19:27.648 "send_buf_size": 4096, 00:19:27.648 "enable_recv_pipe": true, 00:19:27.648 "enable_quickack": false, 00:19:27.648 "enable_placement_id": 0, 00:19:27.648 "enable_zerocopy_send_server": true, 00:19:27.648 "enable_zerocopy_send_client": false, 00:19:27.648 "zerocopy_threshold": 0, 00:19:27.648 "tls_version": 0, 00:19:27.648 "enable_ktls": false 00:19:27.648 } 00:19:27.648 }, 00:19:27.648 { 00:19:27.648 "method": "sock_impl_set_options", 00:19:27.648 "params": { 00:19:27.648 "impl_name": "posix", 00:19:27.648 "recv_buf_size": 2097152, 00:19:27.648 "send_buf_size": 2097152, 00:19:27.648 "enable_recv_pipe": true, 00:19:27.648 "enable_quickack": false, 00:19:27.648 "enable_placement_id": 0, 00:19:27.648 "enable_zerocopy_send_server": true, 00:19:27.648 "enable_zerocopy_send_client": false, 00:19:27.648 "zerocopy_threshold": 0, 00:19:27.648 "tls_version": 0, 00:19:27.648 "enable_ktls": false 00:19:27.648 } 00:19:27.648 } 00:19:27.648 ] 00:19:27.648 }, 00:19:27.648 { 00:19:27.648 "subsystem": "vmd", 00:19:27.648 "config": [] 00:19:27.648 }, 00:19:27.648 { 00:19:27.648 "subsystem": "accel", 00:19:27.648 "config": [ 00:19:27.648 { 00:19:27.648 "method": "accel_set_options", 00:19:27.648 "params": { 00:19:27.648 "small_cache_size": 128, 00:19:27.648 "large_cache_size": 16, 00:19:27.648 "task_count": 2048, 00:19:27.648 "sequence_count": 2048, 00:19:27.648 "buf_count": 2048 00:19:27.648 } 00:19:27.648 } 00:19:27.648 ] 00:19:27.648 }, 00:19:27.648 { 00:19:27.648 "subsystem": "bdev", 00:19:27.648 "config": [ 00:19:27.648 { 00:19:27.648 "method": "bdev_set_options", 00:19:27.648 "params": { 00:19:27.648 "bdev_io_pool_size": 65535, 00:19:27.648 "bdev_io_cache_size": 256, 00:19:27.648 "bdev_auto_examine": true, 00:19:27.648 "iobuf_small_cache_size": 128, 00:19:27.649 "iobuf_large_cache_size": 16 00:19:27.649 } 00:19:27.649 }, 00:19:27.649 { 00:19:27.649 "method": "bdev_raid_set_options", 00:19:27.649 "params": { 00:19:27.649 "process_window_size_kb": 1024 00:19:27.649 } 00:19:27.649 }, 00:19:27.649 { 00:19:27.649 "method": "bdev_iscsi_set_options", 00:19:27.649 "params": { 00:19:27.649 "timeout_sec": 30 00:19:27.649 } 00:19:27.649 }, 00:19:27.649 { 00:19:27.649 "method": "bdev_nvme_set_options", 00:19:27.649 "params": { 00:19:27.649 "action_on_timeout": "none", 00:19:27.649 "timeout_us": 0, 00:19:27.649 "timeout_admin_us": 0, 00:19:27.649 "keep_alive_timeout_ms": 10000, 00:19:27.649 "arbitration_burst": 0, 00:19:27.649 "low_priority_weight": 0, 00:19:27.649 "medium_priority_weight": 0, 00:19:27.649 "high_priority_weight": 0, 00:19:27.649 "nvme_adminq_poll_period_us": 10000, 00:19:27.649 "nvme_ioq_poll_period_us": 0, 00:19:27.649 "io_queue_requests": 512, 00:19:27.649 "delay_cmd_submit": true, 00:19:27.649 "transport_retry_count": 4, 00:19:27.649 "bdev_retry_count": 3, 00:19:27.649 "transport_ack_timeout": 0, 00:19:27.649 "ctrlr_loss_timeout_sec": 0, 00:19:27.649 "reconnect_delay_sec": 0, 00:19:27.649 "fast_io_fail_timeout_sec": 0, 00:19:27.649 "disable_auto_failback": false, 00:19:27.649 "generate_uuids": false, 00:19:27.649 "transport_tos": 0, 00:19:27.649 "nvme_error_stat": false, 00:19:27.649 "rdma_srq_size": 0, 00:19:27.649 "io_path_stat": false, 00:19:27.649 "allow_accel_sequence": false, 00:19:27.649 "rdma_max_cq_size": 0, 00:19:27.649 "rdma_cm_event_timeout_ms": 0, 00:19:27.649 "dhchap_digests": [ 00:19:27.649 "sha256", 00:19:27.649 "sha384", 00:19:27.649 "sha512" 00:19:27.649 ], 00:19:27.649 "dhchap_dhgroups": [ 00:19:27.649 "null", 00:19:27.649 "ffdhe2048", 00:19:27.649 "ffdhe3072", 00:19:27.649 "ffdhe4096", 00:19:27.649 "ffdhe6144", 00:19:27.649 "ffdhe8192" 00:19:27.649 ] 00:19:27.649 } 00:19:27.649 }, 00:19:27.649 { 00:19:27.649 "method": "bdev_nvme_attach_controller", 00:19:27.649 "params": { 00:19:27.649 "name": "nvme0", 00:19:27.649 "trtype": "TCP", 00:19:27.649 "adrfam": "IPv4", 00:19:27.649 "traddr": "10.0.0.2", 00:19:27.649 "trsvcid": "4420", 00:19:27.649 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:19:27.649 "prchk_reftag": false, 00:19:27.649 "prchk_guard": false, 00:19:27.649 "ctrlr_loss_timeout_sec": 0, 00:19:27.649 "reconnect_delay_sec": 0, 00:19:27.649 "fast_io_fail_timeout_sec": 0, 00:19:27.649 "psk": "key0", 00:19:27.649 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:19:27.649 "hdgst": false, 00:19:27.649 "ddgst": false 00:19:27.649 } 00:19:27.649 }, 00:19:27.649 { 00:19:27.649 "method": "bdev_nvme_set_hotplug", 00:19:27.649 "params": { 00:19:27.649 "period_us": 100000, 00:19:27.649 "enable": false 00:19:27.649 } 00:19:27.649 }, 00:19:27.649 { 00:19:27.649 "method": "bdev_enable_histogram", 00:19:27.649 "params": { 00:19:27.649 "name": "nvme0n1", 00:19:27.649 "enable": true 00:19:27.649 } 00:19:27.649 }, 00:19:27.649 { 00:19:27.649 "method": "bdev_wait_for_examine" 00:19:27.649 } 00:19:27.649 ] 00:19:27.649 }, 00:19:27.649 { 00:19:27.649 "subsystem": "nbd", 00:19:27.649 "config": [] 00:19:27.649 } 00:19:27.649 ] 00:19:27.649 }' 00:19:27.649 18:05:21 nvmf_tcp.nvmf_tls -- target/tls.sh@266 -- # killprocess 624344 00:19:27.649 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@948 -- # '[' -z 624344 ']' 00:19:27.649 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@952 -- # kill -0 624344 00:19:27.649 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # uname 00:19:27.649 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:19:27.649 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 624344 00:19:27.908 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:19:27.908 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:19:27.908 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@966 -- # echo 'killing process with pid 624344' 00:19:27.908 killing process with pid 624344 00:19:27.908 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@967 -- # kill 624344 00:19:27.908 Received shutdown signal, test time was about 1.000000 seconds 00:19:27.908 00:19:27.908 Latency(us) 00:19:27.908 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:27.908 =================================================================================================================== 00:19:27.908 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:19:27.908 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@972 -- # wait 624344 00:19:27.908 18:05:21 nvmf_tcp.nvmf_tls -- target/tls.sh@267 -- # killprocess 624102 00:19:27.908 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@948 -- # '[' -z 624102 ']' 00:19:27.908 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@952 -- # kill -0 624102 00:19:27.908 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # uname 00:19:27.908 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:19:27.908 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 624102 00:19:27.908 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:19:27.908 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:19:27.908 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@966 -- # echo 'killing process with pid 624102' 00:19:27.908 killing process with pid 624102 00:19:27.908 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@967 -- # kill 624102 00:19:27.908 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@972 -- # wait 624102 00:19:28.167 18:05:21 nvmf_tcp.nvmf_tls -- target/tls.sh@269 -- # nvmfappstart -c /dev/fd/62 00:19:28.167 18:05:21 nvmf_tcp.nvmf_tls -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:19:28.168 18:05:21 nvmf_tcp.nvmf_tls -- target/tls.sh@269 -- # echo '{ 00:19:28.168 "subsystems": [ 00:19:28.168 { 00:19:28.168 "subsystem": "keyring", 00:19:28.168 "config": [ 00:19:28.168 { 00:19:28.168 "method": "keyring_file_add_key", 00:19:28.168 "params": { 00:19:28.168 "name": "key0", 00:19:28.168 "path": "/tmp/tmp.BYHgc4Ja0k" 00:19:28.168 } 00:19:28.168 } 00:19:28.168 ] 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "subsystem": "iobuf", 00:19:28.168 "config": [ 00:19:28.168 { 00:19:28.168 "method": "iobuf_set_options", 00:19:28.168 "params": { 00:19:28.168 "small_pool_count": 8192, 00:19:28.168 "large_pool_count": 1024, 00:19:28.168 "small_bufsize": 8192, 00:19:28.168 "large_bufsize": 135168 00:19:28.168 } 00:19:28.168 } 00:19:28.168 ] 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "subsystem": "sock", 00:19:28.168 "config": [ 00:19:28.168 { 00:19:28.168 "method": "sock_set_default_impl", 00:19:28.168 "params": { 00:19:28.168 "impl_name": "posix" 00:19:28.168 } 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "method": "sock_impl_set_options", 00:19:28.168 "params": { 00:19:28.168 "impl_name": "ssl", 00:19:28.168 "recv_buf_size": 4096, 00:19:28.168 "send_buf_size": 4096, 00:19:28.168 "enable_recv_pipe": true, 00:19:28.168 "enable_quickack": false, 00:19:28.168 "enable_placement_id": 0, 00:19:28.168 "enable_zerocopy_send_server": true, 00:19:28.168 "enable_zerocopy_send_client": false, 00:19:28.168 "zerocopy_threshold": 0, 00:19:28.168 "tls_version": 0, 00:19:28.168 "enable_ktls": false 00:19:28.168 } 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "method": "sock_impl_set_options", 00:19:28.168 "params": { 00:19:28.168 "impl_name": "posix", 00:19:28.168 "recv_buf_size": 2097152, 00:19:28.168 "send_buf_size": 2097152, 00:19:28.168 "enable_recv_pipe": true, 00:19:28.168 "enable_quickack": false, 00:19:28.168 "enable_placement_id": 0, 00:19:28.168 "enable_zerocopy_send_server": true, 00:19:28.168 "enable_zerocopy_send_client": false, 00:19:28.168 "zerocopy_threshold": 0, 00:19:28.168 "tls_version": 0, 00:19:28.168 "enable_ktls": false 00:19:28.168 } 00:19:28.168 } 00:19:28.168 ] 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "subsystem": "vmd", 00:19:28.168 "config": [] 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "subsystem": "accel", 00:19:28.168 "config": [ 00:19:28.168 { 00:19:28.168 "method": "accel_set_options", 00:19:28.168 "params": { 00:19:28.168 "small_cache_size": 128, 00:19:28.168 "large_cache_size": 16, 00:19:28.168 "task_count": 2048, 00:19:28.168 "sequence_count": 2048, 00:19:28.168 "buf_count": 2048 00:19:28.168 } 00:19:28.168 } 00:19:28.168 ] 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "subsystem": "bdev", 00:19:28.168 "config": [ 00:19:28.168 { 00:19:28.168 "method": "bdev_set_options", 00:19:28.168 "params": { 00:19:28.168 "bdev_io_pool_size": 65535, 00:19:28.168 "bdev_io_cache_size": 256, 00:19:28.168 "bdev_auto_examine": true, 00:19:28.168 "iobuf_small_cache_size": 128, 00:19:28.168 "iobuf_large_cache_size": 16 00:19:28.168 } 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "method": "bdev_raid_set_options", 00:19:28.168 "params": { 00:19:28.168 "process_window_size_kb": 1024 00:19:28.168 } 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "method": "bdev_iscsi_set_options", 00:19:28.168 "params": { 00:19:28.168 "timeout_sec": 30 00:19:28.168 } 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "method": "bdev_nvme_set_options", 00:19:28.168 "params": { 00:19:28.168 "action_on_timeout": "none", 00:19:28.168 "timeout_us": 0, 00:19:28.168 "timeout_admin_us": 0, 00:19:28.168 "keep_alive_timeout_ms": 10000, 00:19:28.168 "arbitration_burst": 0, 00:19:28.168 "low_priority_weight": 0, 00:19:28.168 "medium_priority_weight": 0, 00:19:28.168 "high_priority_weight": 0, 00:19:28.168 "nvme_adminq_poll_period_us": 10000, 00:19:28.168 "nvme_ioq_poll_period_us": 0, 00:19:28.168 "io_queue_requests": 0, 00:19:28.168 "delay_cmd_submit": true, 00:19:28.168 "transport_retry_count": 4, 00:19:28.168 "bdev_retry_count": 3, 00:19:28.168 "transport_ack_timeout": 0, 00:19:28.168 "ctrlr_loss_timeout_sec": 0, 00:19:28.168 "reconnect_delay_sec": 0, 00:19:28.168 "fast_io_fail_timeout_sec": 0, 00:19:28.168 "disable_auto_failback": false, 00:19:28.168 "generate_uuids": false, 00:19:28.168 "transport_tos": 0, 00:19:28.168 "nvme_error_stat": false, 00:19:28.168 "rdma_srq_size": 0, 00:19:28.168 "io_path_stat": false, 00:19:28.168 "allow_accel_sequence": false, 00:19:28.168 "rdma_max_cq_size": 0, 00:19:28.168 "rdma_cm_event_timeout_ms": 0, 00:19:28.168 "dhchap_digests": [ 00:19:28.168 "sha256", 00:19:28.168 "sha384", 00:19:28.168 "sha512" 00:19:28.168 ], 00:19:28.168 "dhchap_dhgroups": [ 00:19:28.168 "null", 00:19:28.168 "ffdhe2048", 00:19:28.168 "ffdhe3072", 00:19:28.168 "ffdhe4096", 00:19:28.168 "ffdhe6144", 00:19:28.168 "ffdhe8192" 00:19:28.168 ] 00:19:28.168 } 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "method": "bdev_nvme_set_hotplug", 00:19:28.168 "params": { 00:19:28.168 "period_us": 100000, 00:19:28.168 "enable": false 00:19:28.168 } 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "method": "bdev_malloc_create", 00:19:28.168 "params": { 00:19:28.168 "name": "malloc0", 00:19:28.168 "num_blocks": 8192, 00:19:28.168 "block_size": 4096, 00:19:28.168 "physical_block_size": 4096, 00:19:28.168 "uuid": "5038965e-4a63-4e67-a174-38221306e7a2", 00:19:28.168 "optimal_io_boundary": 0 00:19:28.168 } 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "method": "bdev_wait_for_examine" 00:19:28.168 } 00:19:28.168 ] 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "subsystem": "nbd", 00:19:28.168 "config": [] 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "subsystem": "scheduler", 00:19:28.168 "config": [ 00:19:28.168 { 00:19:28.168 "method": "framework_set_scheduler", 00:19:28.168 "params": { 00:19:28.168 "name": "static" 00:19:28.168 } 00:19:28.168 } 00:19:28.168 ] 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "subsystem": "nvmf", 00:19:28.168 "config": [ 00:19:28.168 { 00:19:28.168 "method": "nvmf_set_config", 00:19:28.168 "params": { 00:19:28.168 "discovery_filter": "match_any", 00:19:28.168 "admin_cmd_passthru": { 00:19:28.168 "identify_ctrlr": false 00:19:28.168 } 00:19:28.168 } 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "method": "nvmf_set_max_subsystems", 00:19:28.168 "params": { 00:19:28.168 "max_subsystems": 1024 00:19:28.168 } 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "method": "nvmf_set_crdt", 00:19:28.168 "params": { 00:19:28.168 "crdt1": 0, 00:19:28.168 "crdt2": 0, 00:19:28.168 "crdt3": 0 00:19:28.168 } 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "method": "nvmf_create_transport", 00:19:28.168 "params": { 00:19:28.168 "trtype": "TCP", 00:19:28.168 "max_queue_depth": 128, 00:19:28.168 "max_io_qpairs_per_ctrlr": 127, 00:19:28.168 "in_capsule_data_size": 4096, 00:19:28.168 "max_io_size": 131072, 00:19:28.168 "io_unit_size": 131072, 00:19:28.168 "max_aq_depth": 128, 00:19:28.168 "num_shared_buffers": 511, 00:19:28.168 "buf_cache_size": 4294967295, 00:19:28.168 "dif_insert_or_strip": false, 00:19:28.168 "zcopy": false, 00:19:28.168 "c2h_success": false, 00:19:28.168 "sock_priority": 0, 00:19:28.168 "abort_timeout_sec": 1, 00:19:28.168 "ack_timeout": 0, 00:19:28.168 "data_wr_pool_size": 0 00:19:28.168 } 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "method": "nvmf_create_subsystem", 00:19:28.168 "params": { 00:19:28.168 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:28.168 "allow_any_host": false, 00:19:28.168 "serial_number": "00000000000000000000", 00:19:28.168 "model_number": "SPDK bdev Controller", 00:19:28.168 "max_namespaces": 32, 00:19:28.168 "min_cntlid": 1, 00:19:28.168 "max_cntlid": 65519, 00:19:28.168 "ana_reporting": false 00:19:28.168 } 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "method": "nvmf_subsystem_add_host", 00:19:28.168 "params": { 00:19:28.168 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:28.168 "host": "nqn.2016-06.io.spdk:host1", 00:19:28.168 "psk": "key0" 00:19:28.168 } 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "method": "nvmf_subsystem_add_ns", 00:19:28.168 "params": { 00:19:28.168 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:28.168 "namespace": { 00:19:28.168 "nsid": 1, 00:19:28.168 "bdev_name": "malloc0", 00:19:28.168 "nguid": "5038965E4A634E67A17438221306E7A2", 00:19:28.168 "uuid": "5038965e-4a63-4e67-a174-38221306e7a2", 00:19:28.168 "no_auto_visible": false 00:19:28.168 } 00:19:28.168 } 00:19:28.168 }, 00:19:28.168 { 00:19:28.168 "method": "nvmf_subsystem_add_listener", 00:19:28.168 "params": { 00:19:28.168 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:19:28.168 "listen_address": { 00:19:28.168 "trtype": "TCP", 00:19:28.168 "adrfam": "IPv4", 00:19:28.168 "traddr": "10.0.0.2", 00:19:28.168 "trsvcid": "4420" 00:19:28.168 }, 00:19:28.168 "secure_channel": true 00:19:28.168 } 00:19:28.168 } 00:19:28.168 ] 00:19:28.168 } 00:19:28.168 ] 00:19:28.168 }' 00:19:28.169 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@722 -- # xtrace_disable 00:19:28.169 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:28.169 18:05:21 nvmf_tcp.nvmf_tls -- nvmf/common.sh@481 -- # nvmfpid=624826 00:19:28.169 18:05:21 nvmf_tcp.nvmf_tls -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -c /dev/fd/62 00:19:28.169 18:05:21 nvmf_tcp.nvmf_tls -- nvmf/common.sh@482 -- # waitforlisten 624826 00:19:28.169 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@829 -- # '[' -z 624826 ']' 00:19:28.169 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:28.169 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@834 -- # local max_retries=100 00:19:28.169 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:28.169 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:28.169 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@838 -- # xtrace_disable 00:19:28.169 18:05:21 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:28.169 [2024-07-15 18:05:21.844349] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:19:28.169 [2024-07-15 18:05:21.844394] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:28.169 EAL: No free 2048 kB hugepages reported on node 1 00:19:28.428 [2024-07-15 18:05:21.900027] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:28.428 [2024-07-15 18:05:21.978193] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:28.428 [2024-07-15 18:05:21.978233] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:28.428 [2024-07-15 18:05:21.978240] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:28.428 [2024-07-15 18:05:21.978247] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:28.428 [2024-07-15 18:05:21.978252] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:28.428 [2024-07-15 18:05:21.978302] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:19:28.687 [2024-07-15 18:05:22.189220] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:28.687 [2024-07-15 18:05:22.231987] tcp.c: 966:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:28.687 [2024-07-15 18:05:22.232159] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:28.946 18:05:22 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:19:28.946 18:05:22 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@862 -- # return 0 00:19:28.946 18:05:22 nvmf_tcp.nvmf_tls -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:19:28.946 18:05:22 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@728 -- # xtrace_disable 00:19:28.946 18:05:22 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:29.205 18:05:22 nvmf_tcp.nvmf_tls -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:29.205 18:05:22 nvmf_tcp.nvmf_tls -- target/tls.sh@272 -- # bdevperf_pid=625069 00:19:29.205 18:05:22 nvmf_tcp.nvmf_tls -- target/tls.sh@273 -- # waitforlisten 625069 /var/tmp/bdevperf.sock 00:19:29.205 18:05:22 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@829 -- # '[' -z 625069 ']' 00:19:29.205 18:05:22 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:29.205 18:05:22 nvmf_tcp.nvmf_tls -- target/tls.sh@270 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -z -r /var/tmp/bdevperf.sock -q 128 -o 4k -w verify -t 1 -c /dev/fd/63 00:19:29.205 18:05:22 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@834 -- # local max_retries=100 00:19:29.205 18:05:22 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:29.205 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:29.205 18:05:22 nvmf_tcp.nvmf_tls -- target/tls.sh@270 -- # echo '{ 00:19:29.205 "subsystems": [ 00:19:29.205 { 00:19:29.205 "subsystem": "keyring", 00:19:29.205 "config": [ 00:19:29.205 { 00:19:29.205 "method": "keyring_file_add_key", 00:19:29.205 "params": { 00:19:29.205 "name": "key0", 00:19:29.205 "path": "/tmp/tmp.BYHgc4Ja0k" 00:19:29.205 } 00:19:29.205 } 00:19:29.205 ] 00:19:29.205 }, 00:19:29.205 { 00:19:29.205 "subsystem": "iobuf", 00:19:29.205 "config": [ 00:19:29.205 { 00:19:29.205 "method": "iobuf_set_options", 00:19:29.205 "params": { 00:19:29.205 "small_pool_count": 8192, 00:19:29.205 "large_pool_count": 1024, 00:19:29.205 "small_bufsize": 8192, 00:19:29.205 "large_bufsize": 135168 00:19:29.205 } 00:19:29.205 } 00:19:29.205 ] 00:19:29.205 }, 00:19:29.205 { 00:19:29.205 "subsystem": "sock", 00:19:29.205 "config": [ 00:19:29.205 { 00:19:29.205 "method": "sock_set_default_impl", 00:19:29.205 "params": { 00:19:29.205 "impl_name": "posix" 00:19:29.205 } 00:19:29.205 }, 00:19:29.205 { 00:19:29.205 "method": "sock_impl_set_options", 00:19:29.205 "params": { 00:19:29.205 "impl_name": "ssl", 00:19:29.205 "recv_buf_size": 4096, 00:19:29.205 "send_buf_size": 4096, 00:19:29.205 "enable_recv_pipe": true, 00:19:29.205 "enable_quickack": false, 00:19:29.205 "enable_placement_id": 0, 00:19:29.205 "enable_zerocopy_send_server": true, 00:19:29.205 "enable_zerocopy_send_client": false, 00:19:29.205 "zerocopy_threshold": 0, 00:19:29.205 "tls_version": 0, 00:19:29.205 "enable_ktls": false 00:19:29.205 } 00:19:29.205 }, 00:19:29.205 { 00:19:29.205 "method": "sock_impl_set_options", 00:19:29.205 "params": { 00:19:29.205 "impl_name": "posix", 00:19:29.206 "recv_buf_size": 2097152, 00:19:29.206 "send_buf_size": 2097152, 00:19:29.206 "enable_recv_pipe": true, 00:19:29.206 "enable_quickack": false, 00:19:29.206 "enable_placement_id": 0, 00:19:29.206 "enable_zerocopy_send_server": true, 00:19:29.206 "enable_zerocopy_send_client": false, 00:19:29.206 "zerocopy_threshold": 0, 00:19:29.206 "tls_version": 0, 00:19:29.206 "enable_ktls": false 00:19:29.206 } 00:19:29.206 } 00:19:29.206 ] 00:19:29.206 }, 00:19:29.206 { 00:19:29.206 "subsystem": "vmd", 00:19:29.206 "config": [] 00:19:29.206 }, 00:19:29.206 { 00:19:29.206 "subsystem": "accel", 00:19:29.206 "config": [ 00:19:29.206 { 00:19:29.206 "method": "accel_set_options", 00:19:29.206 "params": { 00:19:29.206 "small_cache_size": 128, 00:19:29.206 "large_cache_size": 16, 00:19:29.206 "task_count": 2048, 00:19:29.206 "sequence_count": 2048, 00:19:29.206 "buf_count": 2048 00:19:29.206 } 00:19:29.206 } 00:19:29.206 ] 00:19:29.206 }, 00:19:29.206 { 00:19:29.206 "subsystem": "bdev", 00:19:29.206 "config": [ 00:19:29.206 { 00:19:29.206 "method": "bdev_set_options", 00:19:29.206 "params": { 00:19:29.206 "bdev_io_pool_size": 65535, 00:19:29.206 "bdev_io_cache_size": 256, 00:19:29.206 "bdev_auto_examine": true, 00:19:29.206 "iobuf_small_cache_size": 128, 00:19:29.206 "iobuf_large_cache_size": 16 00:19:29.206 } 00:19:29.206 }, 00:19:29.206 { 00:19:29.206 "method": "bdev_raid_set_options", 00:19:29.206 "params": { 00:19:29.206 "process_window_size_kb": 1024 00:19:29.206 } 00:19:29.206 }, 00:19:29.206 { 00:19:29.206 "method": "bdev_iscsi_set_options", 00:19:29.206 "params": { 00:19:29.206 "timeout_sec": 30 00:19:29.206 } 00:19:29.206 }, 00:19:29.206 { 00:19:29.206 "method": "bdev_nvme_set_options", 00:19:29.206 "params": { 00:19:29.206 "action_on_timeout": "none", 00:19:29.206 "timeout_us": 0, 00:19:29.206 "timeout_admin_us": 0, 00:19:29.206 "keep_alive_timeout_ms": 10000, 00:19:29.206 "arbitration_burst": 0, 00:19:29.206 "low_priority_weight": 0, 00:19:29.206 "medium_priority_weight": 0, 00:19:29.206 "high_priority_weight": 0, 00:19:29.206 "nvme_adminq_poll_period_us": 10000, 00:19:29.206 "nvme_ioq_poll_period_us": 0, 00:19:29.206 "io_queue_requests": 512, 00:19:29.206 "delay_cmd_submit": true, 00:19:29.206 "transport_retry_count": 4, 00:19:29.206 "bdev_retry_count": 3, 00:19:29.206 "transport_ack_timeout": 0, 00:19:29.206 "ctrlr_loss_timeout_sec": 0, 00:19:29.206 "reconnect_delay_sec": 0, 00:19:29.206 "fast_io_fail_timeout_sec": 0, 00:19:29.206 "disable_auto_failback": false, 00:19:29.206 "generate_uuids": false, 00:19:29.206 "transport_tos": 0, 00:19:29.206 "nvme_error_stat": false, 00:19:29.206 "rdma_srq_size": 0, 00:19:29.206 "io_path_stat": false, 00:19:29.206 "allow_accel_sequence": false, 00:19:29.206 "rdma_max_cq_size": 0, 00:19:29.206 "rdma_cm_event_timeout_ms": 0, 00:19:29.206 "dhchap_digests": [ 00:19:29.206 "sha256", 00:19:29.206 "sha384", 00:19:29.206 "sha512" 00:19:29.206 ], 00:19:29.206 "dhchap_dhgroups": [ 00:19:29.206 "null", 00:19:29.206 "ffdhe2048", 00:19:29.206 "ffdhe3072", 00:19:29.206 "ffdhe4096", 00:19:29.206 "ffdhe6144", 00:19:29.206 "ffdhe8192" 00:19:29.206 ] 00:19:29.206 } 00:19:29.206 }, 00:19:29.206 { 00:19:29.206 "method": "bdev_nvme_attach_controller", 00:19:29.206 "params": { 00:19:29.206 "name": "nvme0", 00:19:29.206 "trtype": "TCP", 00:19:29.206 "adrfam": "IPv4", 00:19:29.206 "traddr": "10.0.0.2", 00:19:29.206 "trsvcid": "4420", 00:19:29.206 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:19:29.206 "prchk_reftag": false, 00:19:29.206 "prchk_guard": false, 00:19:29.206 "ctrlr_loss_timeout_sec": 0, 00:19:29.206 "reconnect_delay_sec": 0, 00:19:29.206 "fast_io_fail_timeout_sec": 0, 00:19:29.206 "psk": "key0", 00:19:29.206 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:19:29.206 "hdgst": false, 00:19:29.206 "ddgst": false 00:19:29.206 } 00:19:29.206 }, 00:19:29.206 { 00:19:29.206 "method": "bdev_nvme_set_hotplug", 00:19:29.206 "params": { 00:19:29.206 "period_us": 100000, 00:19:29.206 "enable": false 00:19:29.206 } 00:19:29.206 }, 00:19:29.206 { 00:19:29.206 "method": "bdev_enable_histogram", 00:19:29.206 "params": { 00:19:29.206 "name": "nvme0n1", 00:19:29.206 "enable": true 00:19:29.206 } 00:19:29.206 }, 00:19:29.206 { 00:19:29.206 "method": "bdev_wait_for_examine" 00:19:29.206 } 00:19:29.206 ] 00:19:29.206 }, 00:19:29.206 { 00:19:29.206 "subsystem": "nbd", 00:19:29.206 "config": [] 00:19:29.206 } 00:19:29.206 ] 00:19:29.206 }' 00:19:29.206 18:05:22 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@838 -- # xtrace_disable 00:19:29.206 18:05:22 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:29.206 [2024-07-15 18:05:22.729331] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:19:29.206 [2024-07-15 18:05:22.729379] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid625069 ] 00:19:29.206 EAL: No free 2048 kB hugepages reported on node 1 00:19:29.206 [2024-07-15 18:05:22.783565] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:29.206 [2024-07-15 18:05:22.856884] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:19:29.465 [2024-07-15 18:05:23.008295] bdev_nvme_rpc.c: 517:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:30.043 18:05:23 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:19:30.043 18:05:23 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@862 -- # return 0 00:19:30.043 18:05:23 nvmf_tcp.nvmf_tls -- target/tls.sh@275 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:19:30.043 18:05:23 nvmf_tcp.nvmf_tls -- target/tls.sh@275 -- # jq -r '.[].name' 00:19:30.043 18:05:23 nvmf_tcp.nvmf_tls -- target/tls.sh@275 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:19:30.043 18:05:23 nvmf_tcp.nvmf_tls -- target/tls.sh@276 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:19:30.302 Running I/O for 1 seconds... 00:19:31.236 00:19:31.236 Latency(us) 00:19:31.236 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:31.236 Job: nvme0n1 (Core Mask 0x2, workload: verify, depth: 128, IO size: 4096) 00:19:31.236 Verification LBA range: start 0x0 length 0x2000 00:19:31.236 nvme0n1 : 1.02 4929.42 19.26 0.00 0.00 25730.98 4843.97 62458.66 00:19:31.236 =================================================================================================================== 00:19:31.236 Total : 4929.42 19.26 0.00 0.00 25730.98 4843.97 62458.66 00:19:31.236 0 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- target/tls.sh@278 -- # trap - SIGINT SIGTERM EXIT 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- target/tls.sh@279 -- # cleanup 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- target/tls.sh@15 -- # process_shm --id 0 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@806 -- # type=--id 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@807 -- # id=0 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@808 -- # '[' --id = --pid ']' 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@812 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@812 -- # shm_files=nvmf_trace.0 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@814 -- # [[ -z nvmf_trace.0 ]] 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@818 -- # for n in $shm_files 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@819 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:19:31.236 nvmf_trace.0 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@821 -- # return 0 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- target/tls.sh@16 -- # killprocess 625069 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@948 -- # '[' -z 625069 ']' 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@952 -- # kill -0 625069 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # uname 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 625069 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@966 -- # echo 'killing process with pid 625069' 00:19:31.236 killing process with pid 625069 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@967 -- # kill 625069 00:19:31.236 Received shutdown signal, test time was about 1.000000 seconds 00:19:31.236 00:19:31.236 Latency(us) 00:19:31.236 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:31.236 =================================================================================================================== 00:19:31.236 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:19:31.236 18:05:24 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@972 -- # wait 625069 00:19:31.495 18:05:25 nvmf_tcp.nvmf_tls -- target/tls.sh@17 -- # nvmftestfini 00:19:31.495 18:05:25 nvmf_tcp.nvmf_tls -- nvmf/common.sh@488 -- # nvmfcleanup 00:19:31.495 18:05:25 nvmf_tcp.nvmf_tls -- nvmf/common.sh@117 -- # sync 00:19:31.495 18:05:25 nvmf_tcp.nvmf_tls -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:19:31.495 18:05:25 nvmf_tcp.nvmf_tls -- nvmf/common.sh@120 -- # set +e 00:19:31.495 18:05:25 nvmf_tcp.nvmf_tls -- nvmf/common.sh@121 -- # for i in {1..20} 00:19:31.495 18:05:25 nvmf_tcp.nvmf_tls -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:19:31.495 rmmod nvme_tcp 00:19:31.495 rmmod nvme_fabrics 00:19:31.495 rmmod nvme_keyring 00:19:31.495 18:05:25 nvmf_tcp.nvmf_tls -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:19:31.495 18:05:25 nvmf_tcp.nvmf_tls -- nvmf/common.sh@124 -- # set -e 00:19:31.495 18:05:25 nvmf_tcp.nvmf_tls -- nvmf/common.sh@125 -- # return 0 00:19:31.495 18:05:25 nvmf_tcp.nvmf_tls -- nvmf/common.sh@489 -- # '[' -n 624826 ']' 00:19:31.495 18:05:25 nvmf_tcp.nvmf_tls -- nvmf/common.sh@490 -- # killprocess 624826 00:19:31.495 18:05:25 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@948 -- # '[' -z 624826 ']' 00:19:31.495 18:05:25 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@952 -- # kill -0 624826 00:19:31.495 18:05:25 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # uname 00:19:31.495 18:05:25 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:19:31.495 18:05:25 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 624826 00:19:31.753 18:05:25 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:19:31.753 18:05:25 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:19:31.753 18:05:25 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@966 -- # echo 'killing process with pid 624826' 00:19:31.753 killing process with pid 624826 00:19:31.753 18:05:25 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@967 -- # kill 624826 00:19:31.753 18:05:25 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@972 -- # wait 624826 00:19:31.753 18:05:25 nvmf_tcp.nvmf_tls -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:19:31.754 18:05:25 nvmf_tcp.nvmf_tls -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:19:31.754 18:05:25 nvmf_tcp.nvmf_tls -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:19:31.754 18:05:25 nvmf_tcp.nvmf_tls -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:19:31.754 18:05:25 nvmf_tcp.nvmf_tls -- nvmf/common.sh@278 -- # remove_spdk_ns 00:19:31.754 18:05:25 nvmf_tcp.nvmf_tls -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:19:31.754 18:05:25 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:19:31.754 18:05:25 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:19:34.288 18:05:27 nvmf_tcp.nvmf_tls -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:19:34.288 18:05:27 nvmf_tcp.nvmf_tls -- target/tls.sh@18 -- # rm -f /tmp/tmp.q1a1GxY6ef /tmp/tmp.dBed4Y1irx /tmp/tmp.BYHgc4Ja0k 00:19:34.288 00:19:34.288 real 1m22.828s 00:19:34.288 user 2m6.435s 00:19:34.288 sys 0m29.163s 00:19:34.288 18:05:27 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@1124 -- # xtrace_disable 00:19:34.288 18:05:27 nvmf_tcp.nvmf_tls -- common/autotest_common.sh@10 -- # set +x 00:19:34.288 ************************************ 00:19:34.288 END TEST nvmf_tls 00:19:34.288 ************************************ 00:19:34.288 18:05:27 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:19:34.288 18:05:27 nvmf_tcp -- nvmf/nvmf.sh@62 -- # run_test nvmf_fips /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/fips.sh --transport=tcp 00:19:34.288 18:05:27 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:19:34.288 18:05:27 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:19:34.288 18:05:27 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:19:34.288 ************************************ 00:19:34.288 START TEST nvmf_fips 00:19:34.288 ************************************ 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/fips.sh --transport=tcp 00:19:34.288 * Looking for test storage... 00:19:34.288 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@11 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@7 -- # uname -s 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- paths/export.sh@5 -- # export PATH 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@47 -- # : 0 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@51 -- # have_pci_nics=0 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@12 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@89 -- # check_openssl_version 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@83 -- # local target=3.0.0 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@85 -- # openssl version 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@85 -- # awk '{print $2}' 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@85 -- # ge 3.0.9 3.0.0 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@373 -- # cmp_versions 3.0.9 '>=' 3.0.0 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@330 -- # local ver1 ver1_l 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@331 -- # local ver2 ver2_l 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@333 -- # IFS=.-: 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@333 -- # read -ra ver1 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@334 -- # IFS=.-: 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@334 -- # read -ra ver2 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@335 -- # local 'op=>=' 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@337 -- # ver1_l=3 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@338 -- # ver2_l=3 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@340 -- # local lt=0 gt=0 eq=0 v 00:19:34.288 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@341 -- # case "$op" in 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@345 -- # : 1 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@361 -- # (( v = 0 )) 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@361 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@362 -- # decimal 3 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@350 -- # local d=3 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@351 -- # [[ 3 =~ ^[0-9]+$ ]] 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@352 -- # echo 3 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@362 -- # ver1[v]=3 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@363 -- # decimal 3 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@350 -- # local d=3 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@351 -- # [[ 3 =~ ^[0-9]+$ ]] 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@352 -- # echo 3 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@363 -- # ver2[v]=3 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@364 -- # (( ver1[v] > ver2[v] )) 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@365 -- # (( ver1[v] < ver2[v] )) 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@361 -- # (( v++ )) 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@361 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@362 -- # decimal 0 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@350 -- # local d=0 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@351 -- # [[ 0 =~ ^[0-9]+$ ]] 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@352 -- # echo 0 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@362 -- # ver1[v]=0 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@363 -- # decimal 0 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@350 -- # local d=0 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@351 -- # [[ 0 =~ ^[0-9]+$ ]] 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@352 -- # echo 0 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@363 -- # ver2[v]=0 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@364 -- # (( ver1[v] > ver2[v] )) 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@365 -- # (( ver1[v] < ver2[v] )) 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@361 -- # (( v++ )) 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@361 -- # (( v < (ver1_l > ver2_l ? ver1_l : ver2_l) )) 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@362 -- # decimal 9 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@350 -- # local d=9 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@351 -- # [[ 9 =~ ^[0-9]+$ ]] 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@352 -- # echo 9 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@362 -- # ver1[v]=9 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@363 -- # decimal 0 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@350 -- # local d=0 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@351 -- # [[ 0 =~ ^[0-9]+$ ]] 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@352 -- # echo 0 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@363 -- # ver2[v]=0 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@364 -- # (( ver1[v] > ver2[v] )) 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- scripts/common.sh@364 -- # return 0 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@95 -- # openssl info -modulesdir 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@95 -- # [[ ! -f /usr/lib64/ossl-modules/fips.so ]] 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@100 -- # openssl fipsinstall -help 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@100 -- # warn='This command is not enabled in the Red Hat Enterprise Linux OpenSSL build, please consult Red Hat documentation to learn how to enable FIPS mode' 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@101 -- # [[ This command is not enabled in the Red Hat Enterprise Linux OpenSSL build, please consult Red Hat documentation to learn how to enable FIPS mode == \T\h\i\s\ \c\o\m\m\a\n\d\ \i\s\ \n\o\t\ \e\n\a\b\l\e\d* ]] 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@104 -- # export callback=build_openssl_config 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@104 -- # callback=build_openssl_config 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@113 -- # build_openssl_config 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@37 -- # cat 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@57 -- # [[ ! -t 0 ]] 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@58 -- # cat - 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@114 -- # export OPENSSL_CONF=spdk_fips.conf 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@114 -- # OPENSSL_CONF=spdk_fips.conf 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@116 -- # mapfile -t providers 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@116 -- # openssl list -providers 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@116 -- # grep name 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@120 -- # (( 2 != 2 )) 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@120 -- # [[ name: openssl base provider != *base* ]] 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@120 -- # [[ name: red hat enterprise linux 9 - openssl fips provider != *fips* ]] 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@127 -- # NOT openssl md5 /dev/fd/62 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@127 -- # : 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@648 -- # local es=0 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@650 -- # valid_exec_arg openssl md5 /dev/fd/62 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@636 -- # local arg=openssl 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@640 -- # type -t openssl 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@642 -- # type -P openssl 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@642 -- # arg=/usr/bin/openssl 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@642 -- # [[ -x /usr/bin/openssl ]] 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@651 -- # openssl md5 /dev/fd/62 00:19:34.289 Error setting digest 00:19:34.289 0002691DD67F0000:error:0308010C:digital envelope routines:inner_evp_generic_fetch:unsupported:crypto/evp/evp_fetch.c:373:Global default library context, Algorithm (MD5 : 97), Properties () 00:19:34.289 0002691DD67F0000:error:03000086:digital envelope routines:evp_md_init_internal:initialization error:crypto/evp/digest.c:254: 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@651 -- # es=1 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- fips/fips.sh@130 -- # nvmftestinit 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@448 -- # prepare_net_devs 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@410 -- # local -g is_hw=no 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@412 -- # remove_spdk_ns 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- nvmf/common.sh@285 -- # xtrace_disable 00:19:34.289 18:05:27 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@291 -- # pci_devs=() 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@291 -- # local -a pci_devs 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@292 -- # pci_net_devs=() 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@293 -- # pci_drivers=() 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@293 -- # local -A pci_drivers 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@295 -- # net_devs=() 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@295 -- # local -ga net_devs 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@296 -- # e810=() 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@296 -- # local -ga e810 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@297 -- # x722=() 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@297 -- # local -ga x722 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@298 -- # mlx=() 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@298 -- # local -ga mlx 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:19:39.559 Found 0000:86:00.0 (0x8086 - 0x159b) 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:19:39.559 Found 0000:86:00.1 (0x8086 - 0x159b) 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@390 -- # [[ up == up ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:19:39.559 Found net devices under 0000:86:00.0: cvl_0_0 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@390 -- # [[ up == up ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:19:39.559 Found net devices under 0000:86:00.1: cvl_0_1 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@414 -- # is_hw=yes 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:19:39.559 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:19:39.560 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:19:39.560 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:19:39.560 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:19:39.560 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:19:39.560 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:19:39.560 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:19:39.560 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:19:39.560 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:19:39.560 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:19:39.560 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:19:39.560 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:19:39.560 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:19:39.560 18:05:32 nvmf_tcp.nvmf_fips -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:19:39.560 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:19:39.560 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.176 ms 00:19:39.560 00:19:39.560 --- 10.0.0.2 ping statistics --- 00:19:39.560 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:19:39.560 rtt min/avg/max/mdev = 0.176/0.176/0.176/0.000 ms 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:19:39.560 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:19:39.560 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.244 ms 00:19:39.560 00:19:39.560 --- 10.0.0.1 ping statistics --- 00:19:39.560 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:19:39.560 rtt min/avg/max/mdev = 0.244/0.244/0.244/0.000 ms 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- nvmf/common.sh@422 -- # return 0 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- fips/fips.sh@131 -- # nvmfappstart -m 0x2 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@722 -- # xtrace_disable 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- nvmf/common.sh@481 -- # nvmfpid=628864 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- nvmf/common.sh@482 -- # waitforlisten 628864 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@829 -- # '[' -z 628864 ']' 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@834 -- # local max_retries=100 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:19:39.560 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@838 -- # xtrace_disable 00:19:39.560 18:05:33 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:19:39.560 [2024-07-15 18:05:33.161871] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:19:39.560 [2024-07-15 18:05:33.161918] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:19:39.560 EAL: No free 2048 kB hugepages reported on node 1 00:19:39.560 [2024-07-15 18:05:33.216987] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:39.818 [2024-07-15 18:05:33.295526] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:19:39.818 [2024-07-15 18:05:33.295560] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:19:39.818 [2024-07-15 18:05:33.295567] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:19:39.818 [2024-07-15 18:05:33.295572] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:19:39.818 [2024-07-15 18:05:33.295577] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:19:39.818 [2024-07-15 18:05:33.295593] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:19:40.384 18:05:33 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:19:40.384 18:05:33 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@862 -- # return 0 00:19:40.384 18:05:33 nvmf_tcp.nvmf_fips -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:19:40.384 18:05:33 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@728 -- # xtrace_disable 00:19:40.384 18:05:33 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:19:40.384 18:05:33 nvmf_tcp.nvmf_fips -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:19:40.384 18:05:33 nvmf_tcp.nvmf_fips -- fips/fips.sh@133 -- # trap cleanup EXIT 00:19:40.384 18:05:33 nvmf_tcp.nvmf_fips -- fips/fips.sh@136 -- # key=NVMeTLSkey-1:01:VRLbtnN9AQb2WXW3c9+wEf/DRLz0QuLdbYvEhwtdWwNf9LrZ: 00:19:40.384 18:05:33 nvmf_tcp.nvmf_fips -- fips/fips.sh@137 -- # key_path=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/key.txt 00:19:40.384 18:05:33 nvmf_tcp.nvmf_fips -- fips/fips.sh@138 -- # echo -n NVMeTLSkey-1:01:VRLbtnN9AQb2WXW3c9+wEf/DRLz0QuLdbYvEhwtdWwNf9LrZ: 00:19:40.384 18:05:33 nvmf_tcp.nvmf_fips -- fips/fips.sh@139 -- # chmod 0600 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/key.txt 00:19:40.384 18:05:33 nvmf_tcp.nvmf_fips -- fips/fips.sh@141 -- # setup_nvmf_tgt_conf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/key.txt 00:19:40.384 18:05:33 nvmf_tcp.nvmf_fips -- fips/fips.sh@22 -- # local key=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/key.txt 00:19:40.384 18:05:33 nvmf_tcp.nvmf_fips -- fips/fips.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:19:40.642 [2024-07-15 18:05:34.139058] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:19:40.643 [2024-07-15 18:05:34.155062] tcp.c: 966:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:19:40.643 [2024-07-15 18:05:34.155221] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:19:40.643 [2024-07-15 18:05:34.183334] tcp.c:3710:nvmf_tcp_subsystem_add_host: *WARNING*: nvmf_tcp_psk_path: deprecated feature PSK path to be removed in v24.09 00:19:40.643 malloc0 00:19:40.643 18:05:34 nvmf_tcp.nvmf_fips -- fips/fips.sh@144 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:19:40.643 18:05:34 nvmf_tcp.nvmf_fips -- fips/fips.sh@147 -- # bdevperf_pid=629113 00:19:40.643 18:05:34 nvmf_tcp.nvmf_fips -- fips/fips.sh@148 -- # waitforlisten 629113 /var/tmp/bdevperf.sock 00:19:40.643 18:05:34 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@829 -- # '[' -z 629113 ']' 00:19:40.643 18:05:34 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:19:40.643 18:05:34 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@834 -- # local max_retries=100 00:19:40.643 18:05:34 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:19:40.643 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:19:40.643 18:05:34 nvmf_tcp.nvmf_fips -- fips/fips.sh@145 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 10 00:19:40.643 18:05:34 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@838 -- # xtrace_disable 00:19:40.643 18:05:34 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:19:40.643 [2024-07-15 18:05:34.271716] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:19:40.643 [2024-07-15 18:05:34.271762] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid629113 ] 00:19:40.643 EAL: No free 2048 kB hugepages reported on node 1 00:19:40.643 [2024-07-15 18:05:34.321802] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:19:40.902 [2024-07-15 18:05:34.401141] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:19:41.539 18:05:35 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:19:41.539 18:05:35 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@862 -- # return 0 00:19:41.539 18:05:35 nvmf_tcp.nvmf_fips -- fips/fips.sh@150 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b TLSTEST -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -q nqn.2016-06.io.spdk:host1 --psk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/key.txt 00:19:41.539 [2024-07-15 18:05:35.222831] bdev_nvme_rpc.c: 517:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:19:41.539 [2024-07-15 18:05:35.222903] nvme_tcp.c:2589:nvme_tcp_generate_tls_credentials: *WARNING*: nvme_ctrlr_psk: deprecated feature spdk_nvme_ctrlr_opts.psk to be removed in v24.09 00:19:41.797 TLSTESTn1 00:19:41.798 18:05:35 nvmf_tcp.nvmf_fips -- fips/fips.sh@154 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:19:41.798 Running I/O for 10 seconds... 00:19:51.774 00:19:51.774 Latency(us) 00:19:51.774 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:51.774 Job: TLSTESTn1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:19:51.774 Verification LBA range: start 0x0 length 0x2000 00:19:51.774 TLSTESTn1 : 10.01 5471.17 21.37 0.00 0.00 23357.96 4872.46 31229.33 00:19:51.774 =================================================================================================================== 00:19:51.774 Total : 5471.17 21.37 0.00 0.00 23357.96 4872.46 31229.33 00:19:51.774 0 00:19:51.774 18:05:45 nvmf_tcp.nvmf_fips -- fips/fips.sh@1 -- # cleanup 00:19:51.774 18:05:45 nvmf_tcp.nvmf_fips -- fips/fips.sh@15 -- # process_shm --id 0 00:19:51.774 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@806 -- # type=--id 00:19:51.774 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@807 -- # id=0 00:19:51.774 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@808 -- # '[' --id = --pid ']' 00:19:51.774 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@812 -- # find /dev/shm -name '*.0' -printf '%f\n' 00:19:51.774 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@812 -- # shm_files=nvmf_trace.0 00:19:51.774 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@814 -- # [[ -z nvmf_trace.0 ]] 00:19:51.774 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@818 -- # for n in $shm_files 00:19:51.774 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@819 -- # tar -C /dev/shm/ -cvzf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvmf_trace.0_shm.tar.gz nvmf_trace.0 00:19:51.774 nvmf_trace.0 00:19:52.033 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@821 -- # return 0 00:19:52.033 18:05:45 nvmf_tcp.nvmf_fips -- fips/fips.sh@16 -- # killprocess 629113 00:19:52.033 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@948 -- # '[' -z 629113 ']' 00:19:52.033 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@952 -- # kill -0 629113 00:19:52.033 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@953 -- # uname 00:19:52.033 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:19:52.033 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 629113 00:19:52.033 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@954 -- # process_name=reactor_2 00:19:52.033 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@958 -- # '[' reactor_2 = sudo ']' 00:19:52.033 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@966 -- # echo 'killing process with pid 629113' 00:19:52.033 killing process with pid 629113 00:19:52.033 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@967 -- # kill 629113 00:19:52.033 Received shutdown signal, test time was about 10.000000 seconds 00:19:52.033 00:19:52.033 Latency(us) 00:19:52.033 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:19:52.033 =================================================================================================================== 00:19:52.033 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:19:52.033 [2024-07-15 18:05:45.565753] app.c:1024:log_deprecation_hits: *WARNING*: nvme_ctrlr_psk: deprecation 'spdk_nvme_ctrlr_opts.psk' scheduled for removal in v24.09 hit 1 times 00:19:52.033 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@972 -- # wait 629113 00:19:52.033 18:05:45 nvmf_tcp.nvmf_fips -- fips/fips.sh@17 -- # nvmftestfini 00:19:52.033 18:05:45 nvmf_tcp.nvmf_fips -- nvmf/common.sh@488 -- # nvmfcleanup 00:19:52.033 18:05:45 nvmf_tcp.nvmf_fips -- nvmf/common.sh@117 -- # sync 00:19:52.033 18:05:45 nvmf_tcp.nvmf_fips -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:19:52.033 18:05:45 nvmf_tcp.nvmf_fips -- nvmf/common.sh@120 -- # set +e 00:19:52.033 18:05:45 nvmf_tcp.nvmf_fips -- nvmf/common.sh@121 -- # for i in {1..20} 00:19:52.033 18:05:45 nvmf_tcp.nvmf_fips -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:19:52.033 rmmod nvme_tcp 00:19:52.295 rmmod nvme_fabrics 00:19:52.295 rmmod nvme_keyring 00:19:52.295 18:05:45 nvmf_tcp.nvmf_fips -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:19:52.295 18:05:45 nvmf_tcp.nvmf_fips -- nvmf/common.sh@124 -- # set -e 00:19:52.295 18:05:45 nvmf_tcp.nvmf_fips -- nvmf/common.sh@125 -- # return 0 00:19:52.295 18:05:45 nvmf_tcp.nvmf_fips -- nvmf/common.sh@489 -- # '[' -n 628864 ']' 00:19:52.295 18:05:45 nvmf_tcp.nvmf_fips -- nvmf/common.sh@490 -- # killprocess 628864 00:19:52.295 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@948 -- # '[' -z 628864 ']' 00:19:52.295 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@952 -- # kill -0 628864 00:19:52.295 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@953 -- # uname 00:19:52.295 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:19:52.295 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 628864 00:19:52.295 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:19:52.295 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:19:52.295 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@966 -- # echo 'killing process with pid 628864' 00:19:52.295 killing process with pid 628864 00:19:52.295 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@967 -- # kill 628864 00:19:52.295 [2024-07-15 18:05:45.844122] app.c:1024:log_deprecation_hits: *WARNING*: nvmf_tcp_psk_path: deprecation 'PSK path' scheduled for removal in v24.09 hit 1 times 00:19:52.295 18:05:45 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@972 -- # wait 628864 00:19:52.554 18:05:46 nvmf_tcp.nvmf_fips -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:19:52.554 18:05:46 nvmf_tcp.nvmf_fips -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:19:52.554 18:05:46 nvmf_tcp.nvmf_fips -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:19:52.554 18:05:46 nvmf_tcp.nvmf_fips -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:19:52.554 18:05:46 nvmf_tcp.nvmf_fips -- nvmf/common.sh@278 -- # remove_spdk_ns 00:19:52.554 18:05:46 nvmf_tcp.nvmf_fips -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:19:52.554 18:05:46 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:19:52.554 18:05:46 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:19:54.460 18:05:48 nvmf_tcp.nvmf_fips -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:19:54.460 18:05:48 nvmf_tcp.nvmf_fips -- fips/fips.sh@18 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/fips/key.txt 00:19:54.460 00:19:54.460 real 0m20.521s 00:19:54.460 user 0m22.816s 00:19:54.460 sys 0m8.482s 00:19:54.460 18:05:48 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@1124 -- # xtrace_disable 00:19:54.460 18:05:48 nvmf_tcp.nvmf_fips -- common/autotest_common.sh@10 -- # set +x 00:19:54.460 ************************************ 00:19:54.460 END TEST nvmf_fips 00:19:54.460 ************************************ 00:19:54.460 18:05:48 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:19:54.460 18:05:48 nvmf_tcp -- nvmf/nvmf.sh@65 -- # '[' 0 -eq 1 ']' 00:19:54.460 18:05:48 nvmf_tcp -- nvmf/nvmf.sh@71 -- # [[ phy == phy ]] 00:19:54.460 18:05:48 nvmf_tcp -- nvmf/nvmf.sh@72 -- # '[' tcp = tcp ']' 00:19:54.460 18:05:48 nvmf_tcp -- nvmf/nvmf.sh@73 -- # gather_supported_nvmf_pci_devs 00:19:54.460 18:05:48 nvmf_tcp -- nvmf/common.sh@285 -- # xtrace_disable 00:19:54.460 18:05:48 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@291 -- # pci_devs=() 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@291 -- # local -a pci_devs 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@292 -- # pci_net_devs=() 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@293 -- # pci_drivers=() 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@293 -- # local -A pci_drivers 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@295 -- # net_devs=() 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@295 -- # local -ga net_devs 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@296 -- # e810=() 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@296 -- # local -ga e810 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@297 -- # x722=() 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@297 -- # local -ga x722 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@298 -- # mlx=() 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@298 -- # local -ga mlx 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:19:59.739 Found 0000:86:00.0 (0x8086 - 0x159b) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:19:59.739 Found 0000:86:00.1 (0x8086 - 0x159b) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@390 -- # [[ up == up ]] 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:19:59.739 Found net devices under 0000:86:00.0: cvl_0_0 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@390 -- # [[ up == up ]] 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:19:59.739 Found net devices under 0000:86:00.1: cvl_0_1 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/nvmf.sh@74 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/nvmf.sh@75 -- # (( 2 > 0 )) 00:19:59.739 18:05:53 nvmf_tcp -- nvmf/nvmf.sh@76 -- # run_test nvmf_perf_adq /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/perf_adq.sh --transport=tcp 00:19:59.739 18:05:53 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:19:59.739 18:05:53 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:19:59.740 18:05:53 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:19:59.740 ************************************ 00:19:59.740 START TEST nvmf_perf_adq 00:19:59.740 ************************************ 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/perf_adq.sh --transport=tcp 00:19:59.740 * Looking for test storage... 00:19:59.740 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@7 -- # uname -s 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- paths/export.sh@5 -- # export PATH 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:19:59.740 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@47 -- # : 0 00:19:59.998 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:19:59.998 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:19:59.998 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:19:59.998 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:19:59.998 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:19:59.998 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:19:59.998 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:19:59.998 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@51 -- # have_pci_nics=0 00:19:59.998 18:05:53 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@11 -- # gather_supported_nvmf_pci_devs 00:19:59.998 18:05:53 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@285 -- # xtrace_disable 00:19:59.998 18:05:53 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@291 -- # pci_devs=() 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@291 -- # local -a pci_devs 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@292 -- # pci_net_devs=() 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@293 -- # pci_drivers=() 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@293 -- # local -A pci_drivers 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@295 -- # net_devs=() 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@295 -- # local -ga net_devs 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@296 -- # e810=() 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@296 -- # local -ga e810 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@297 -- # x722=() 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@297 -- # local -ga x722 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@298 -- # mlx=() 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@298 -- # local -ga mlx 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:20:05.266 Found 0000:86:00.0 (0x8086 - 0x159b) 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:20:05.266 Found 0000:86:00.1 (0x8086 - 0x159b) 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@390 -- # [[ up == up ]] 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:20:05.266 Found net devices under 0000:86:00.0: cvl_0_0 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:20:05.266 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:05.267 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:20:05.267 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:05.267 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@390 -- # [[ up == up ]] 00:20:05.267 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:20:05.267 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:05.267 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:20:05.267 Found net devices under 0000:86:00.1: cvl_0_1 00:20:05.267 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:20:05.267 18:05:58 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:20:05.267 18:05:58 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@12 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:20:05.267 18:05:58 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@13 -- # (( 2 == 0 )) 00:20:05.267 18:05:58 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@18 -- # perf=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf 00:20:05.267 18:05:58 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@60 -- # adq_reload_driver 00:20:05.267 18:05:58 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@53 -- # rmmod ice 00:20:06.644 18:05:59 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@54 -- # modprobe ice 00:20:08.550 18:06:01 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@55 -- # sleep 5 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@68 -- # nvmftestinit 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@448 -- # prepare_net_devs 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@410 -- # local -g is_hw=no 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@412 -- # remove_spdk_ns 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@285 -- # xtrace_disable 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@291 -- # pci_devs=() 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@291 -- # local -a pci_devs 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@292 -- # pci_net_devs=() 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@293 -- # pci_drivers=() 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@293 -- # local -A pci_drivers 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@295 -- # net_devs=() 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@295 -- # local -ga net_devs 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@296 -- # e810=() 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@296 -- # local -ga e810 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@297 -- # x722=() 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@297 -- # local -ga x722 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@298 -- # mlx=() 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@298 -- # local -ga mlx 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:20:13.885 Found 0000:86:00.0 (0x8086 - 0x159b) 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:20:13.885 Found 0000:86:00.1 (0x8086 - 0x159b) 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@390 -- # [[ up == up ]] 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:20:13.885 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:20:13.886 Found net devices under 0000:86:00.0: cvl_0_0 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@390 -- # [[ up == up ]] 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:20:13.886 Found net devices under 0000:86:00.1: cvl_0_1 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@414 -- # is_hw=yes 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:20:13.886 18:06:06 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:20:13.886 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:20:13.886 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.151 ms 00:20:13.886 00:20:13.886 --- 10.0.0.2 ping statistics --- 00:20:13.886 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:13.886 rtt min/avg/max/mdev = 0.151/0.151/0.151/0.000 ms 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:20:13.886 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:20:13.886 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.203 ms 00:20:13.886 00:20:13.886 --- 10.0.0.1 ping statistics --- 00:20:13.886 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:13.886 rtt min/avg/max/mdev = 0.203/0.203/0.203/0.000 ms 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@422 -- # return 0 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@69 -- # nvmfappstart -m 0xF --wait-for-rpc 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@722 -- # xtrace_disable 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@481 -- # nvmfpid=638948 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@482 -- # waitforlisten 638948 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@829 -- # '[' -z 638948 ']' 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@834 -- # local max_retries=100 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:13.886 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@838 -- # xtrace_disable 00:20:13.886 18:06:07 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:13.886 [2024-07-15 18:06:07.223786] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:20:13.886 [2024-07-15 18:06:07.223825] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:13.886 EAL: No free 2048 kB hugepages reported on node 1 00:20:13.886 [2024-07-15 18:06:07.279582] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:20:13.886 [2024-07-15 18:06:07.359376] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:20:13.886 [2024-07-15 18:06:07.359414] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:20:13.886 [2024-07-15 18:06:07.359421] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:20:13.886 [2024-07-15 18:06:07.359427] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:20:13.886 [2024-07-15 18:06:07.359432] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:20:13.886 [2024-07-15 18:06:07.359474] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:20:13.886 [2024-07-15 18:06:07.359551] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:20:13.886 [2024-07-15 18:06:07.359629] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:20:13.886 [2024-07-15 18:06:07.359630] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:20:14.453 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:20:14.453 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@862 -- # return 0 00:20:14.453 18:06:08 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:20:14.453 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@728 -- # xtrace_disable 00:20:14.453 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:14.453 18:06:08 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:20:14.453 18:06:08 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@70 -- # adq_configure_nvmf_target 0 00:20:14.453 18:06:08 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@42 -- # rpc_cmd sock_get_default_impl 00:20:14.453 18:06:08 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@42 -- # jq -r .impl_name 00:20:14.453 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:14.453 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:14.453 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:14.453 18:06:08 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@42 -- # socket_impl=posix 00:20:14.453 18:06:08 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@43 -- # rpc_cmd sock_impl_set_options --enable-placement-id 0 --enable-zerocopy-send-server -i posix 00:20:14.453 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:14.453 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:14.453 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:14.453 18:06:08 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@44 -- # rpc_cmd framework_start_init 00:20:14.453 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:14.453 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:14.711 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@45 -- # rpc_cmd nvmf_create_transport -t tcp -o --io-unit-size 8192 --sock-priority 0 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:14.712 [2024-07-15 18:06:08.212664] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@46 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:14.712 Malloc1 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@47 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@48 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@49 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:14.712 [2024-07-15 18:06:08.260640] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@74 -- # perfpid=639322 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@75 -- # sleep 2 00:20:14.712 18:06:08 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@71 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 64 -o 4096 -w randread -t 10 -c 0xF0 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:20:14.712 EAL: No free 2048 kB hugepages reported on node 1 00:20:16.616 18:06:10 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@77 -- # rpc_cmd nvmf_get_stats 00:20:16.616 18:06:10 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:16.616 18:06:10 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:16.616 18:06:10 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:16.616 18:06:10 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@77 -- # nvmf_stats='{ 00:20:16.616 "tick_rate": 2300000000, 00:20:16.616 "poll_groups": [ 00:20:16.616 { 00:20:16.616 "name": "nvmf_tgt_poll_group_000", 00:20:16.616 "admin_qpairs": 1, 00:20:16.616 "io_qpairs": 1, 00:20:16.616 "current_admin_qpairs": 1, 00:20:16.616 "current_io_qpairs": 1, 00:20:16.616 "pending_bdev_io": 0, 00:20:16.616 "completed_nvme_io": 20482, 00:20:16.616 "transports": [ 00:20:16.616 { 00:20:16.616 "trtype": "TCP" 00:20:16.616 } 00:20:16.616 ] 00:20:16.616 }, 00:20:16.616 { 00:20:16.616 "name": "nvmf_tgt_poll_group_001", 00:20:16.616 "admin_qpairs": 0, 00:20:16.616 "io_qpairs": 1, 00:20:16.616 "current_admin_qpairs": 0, 00:20:16.616 "current_io_qpairs": 1, 00:20:16.616 "pending_bdev_io": 0, 00:20:16.616 "completed_nvme_io": 20954, 00:20:16.616 "transports": [ 00:20:16.616 { 00:20:16.616 "trtype": "TCP" 00:20:16.616 } 00:20:16.616 ] 00:20:16.616 }, 00:20:16.616 { 00:20:16.616 "name": "nvmf_tgt_poll_group_002", 00:20:16.616 "admin_qpairs": 0, 00:20:16.616 "io_qpairs": 1, 00:20:16.616 "current_admin_qpairs": 0, 00:20:16.616 "current_io_qpairs": 1, 00:20:16.616 "pending_bdev_io": 0, 00:20:16.616 "completed_nvme_io": 20783, 00:20:16.616 "transports": [ 00:20:16.616 { 00:20:16.616 "trtype": "TCP" 00:20:16.616 } 00:20:16.616 ] 00:20:16.616 }, 00:20:16.616 { 00:20:16.616 "name": "nvmf_tgt_poll_group_003", 00:20:16.616 "admin_qpairs": 0, 00:20:16.616 "io_qpairs": 1, 00:20:16.616 "current_admin_qpairs": 0, 00:20:16.616 "current_io_qpairs": 1, 00:20:16.616 "pending_bdev_io": 0, 00:20:16.616 "completed_nvme_io": 20428, 00:20:16.616 "transports": [ 00:20:16.616 { 00:20:16.616 "trtype": "TCP" 00:20:16.616 } 00:20:16.616 ] 00:20:16.616 } 00:20:16.616 ] 00:20:16.616 }' 00:20:16.616 18:06:10 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@78 -- # jq -r '.poll_groups[] | select(.current_io_qpairs == 1) | length' 00:20:16.616 18:06:10 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@78 -- # wc -l 00:20:16.616 18:06:10 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@78 -- # count=4 00:20:16.616 18:06:10 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@79 -- # [[ 4 -ne 4 ]] 00:20:16.616 18:06:10 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@83 -- # wait 639322 00:20:24.730 Initializing NVMe Controllers 00:20:24.730 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:20:24.730 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 4 00:20:24.730 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 5 00:20:24.730 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 6 00:20:24.730 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 7 00:20:24.730 Initialization complete. Launching workers. 00:20:24.730 ======================================================== 00:20:24.730 Latency(us) 00:20:24.730 Device Information : IOPS MiB/s Average min max 00:20:24.730 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 4: 10518.30 41.09 6084.42 1574.83 11136.04 00:20:24.730 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 5: 10845.40 42.36 5901.73 1993.07 11182.50 00:20:24.730 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 6: 10734.50 41.93 5962.10 1748.49 9520.88 00:20:24.730 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 7: 10644.70 41.58 6012.70 1882.10 10029.86 00:20:24.730 ======================================================== 00:20:24.730 Total : 42742.89 166.96 5989.48 1574.83 11182.50 00:20:24.730 00:20:24.730 18:06:18 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@84 -- # nvmftestfini 00:20:24.730 18:06:18 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@488 -- # nvmfcleanup 00:20:24.730 18:06:18 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@117 -- # sync 00:20:24.730 18:06:18 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:20:24.730 18:06:18 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@120 -- # set +e 00:20:24.730 18:06:18 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@121 -- # for i in {1..20} 00:20:24.730 18:06:18 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:20:24.730 rmmod nvme_tcp 00:20:24.730 rmmod nvme_fabrics 00:20:24.730 rmmod nvme_keyring 00:20:24.730 18:06:18 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:20:24.730 18:06:18 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@124 -- # set -e 00:20:24.730 18:06:18 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@125 -- # return 0 00:20:24.730 18:06:18 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@489 -- # '[' -n 638948 ']' 00:20:24.730 18:06:18 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@490 -- # killprocess 638948 00:20:24.730 18:06:18 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@948 -- # '[' -z 638948 ']' 00:20:24.730 18:06:18 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@952 -- # kill -0 638948 00:20:24.730 18:06:18 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@953 -- # uname 00:20:24.730 18:06:18 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:20:24.730 18:06:18 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 638948 00:20:24.989 18:06:18 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:20:24.989 18:06:18 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:20:24.989 18:06:18 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@966 -- # echo 'killing process with pid 638948' 00:20:24.989 killing process with pid 638948 00:20:24.989 18:06:18 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@967 -- # kill 638948 00:20:24.989 18:06:18 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@972 -- # wait 638948 00:20:24.989 18:06:18 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:20:24.989 18:06:18 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:20:24.989 18:06:18 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:20:24.989 18:06:18 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:20:24.989 18:06:18 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@278 -- # remove_spdk_ns 00:20:24.989 18:06:18 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:24.989 18:06:18 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:20:24.989 18:06:18 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:27.522 18:06:20 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:20:27.522 18:06:20 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@86 -- # adq_reload_driver 00:20:27.522 18:06:20 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@53 -- # rmmod ice 00:20:28.458 18:06:21 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@54 -- # modprobe ice 00:20:30.356 18:06:23 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@55 -- # sleep 5 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@89 -- # nvmftestinit 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@448 -- # prepare_net_devs 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@410 -- # local -g is_hw=no 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@412 -- # remove_spdk_ns 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@285 -- # xtrace_disable 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@291 -- # pci_devs=() 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@291 -- # local -a pci_devs 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@292 -- # pci_net_devs=() 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@293 -- # pci_drivers=() 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@293 -- # local -A pci_drivers 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@295 -- # net_devs=() 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@295 -- # local -ga net_devs 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@296 -- # e810=() 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@296 -- # local -ga e810 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@297 -- # x722=() 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@297 -- # local -ga x722 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@298 -- # mlx=() 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@298 -- # local -ga mlx 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:20:35.632 Found 0000:86:00.0 (0x8086 - 0x159b) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:20:35.632 Found 0000:86:00.1 (0x8086 - 0x159b) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@390 -- # [[ up == up ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:20:35.632 Found net devices under 0000:86:00.0: cvl_0_0 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@390 -- # [[ up == up ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:20:35.632 Found net devices under 0000:86:00.1: cvl_0_1 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@414 -- # is_hw=yes 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:20:35.632 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:20:35.633 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:20:35.633 18:06:28 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:20:35.633 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:20:35.633 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.179 ms 00:20:35.633 00:20:35.633 --- 10.0.0.2 ping statistics --- 00:20:35.633 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:35.633 rtt min/avg/max/mdev = 0.179/0.179/0.179/0.000 ms 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:20:35.633 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:20:35.633 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.178 ms 00:20:35.633 00:20:35.633 --- 10.0.0.1 ping statistics --- 00:20:35.633 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:35.633 rtt min/avg/max/mdev = 0.178/0.178/0.178/0.000 ms 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@422 -- # return 0 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@90 -- # adq_configure_driver 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@22 -- # ip netns exec cvl_0_0_ns_spdk ethtool --offload cvl_0_0 hw-tc-offload on 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@24 -- # ip netns exec cvl_0_0_ns_spdk ethtool --set-priv-flags cvl_0_0 channel-pkt-inspect-optimize off 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@26 -- # sysctl -w net.core.busy_poll=1 00:20:35.633 net.core.busy_poll = 1 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@27 -- # sysctl -w net.core.busy_read=1 00:20:35.633 net.core.busy_read = 1 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@29 -- # tc=/usr/sbin/tc 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@31 -- # ip netns exec cvl_0_0_ns_spdk /usr/sbin/tc qdisc add dev cvl_0_0 root mqprio num_tc 2 map 0 1 queues 2@0 2@2 hw 1 mode channel 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@33 -- # ip netns exec cvl_0_0_ns_spdk /usr/sbin/tc qdisc add dev cvl_0_0 ingress 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@35 -- # ip netns exec cvl_0_0_ns_spdk /usr/sbin/tc filter add dev cvl_0_0 protocol ip parent ffff: prio 1 flower dst_ip 10.0.0.2/32 ip_proto tcp dst_port 4420 skip_sw hw_tc 1 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@38 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/nvmf/set_xps_rxqs cvl_0_0 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@91 -- # nvmfappstart -m 0xF --wait-for-rpc 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@722 -- # xtrace_disable 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@481 -- # nvmfpid=643371 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@482 -- # waitforlisten 643371 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@829 -- # '[' -z 643371 ']' 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@834 -- # local max_retries=100 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:35.633 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@838 -- # xtrace_disable 00:20:35.633 18:06:29 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:35.890 [2024-07-15 18:06:29.405047] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:20:35.890 [2024-07-15 18:06:29.405097] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:35.890 EAL: No free 2048 kB hugepages reported on node 1 00:20:35.890 [2024-07-15 18:06:29.464218] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:20:35.890 [2024-07-15 18:06:29.540614] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:20:35.890 [2024-07-15 18:06:29.540655] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:20:35.890 [2024-07-15 18:06:29.540662] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:20:35.890 [2024-07-15 18:06:29.540671] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:20:35.890 [2024-07-15 18:06:29.540676] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:20:35.890 [2024-07-15 18:06:29.540773] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:20:35.890 [2024-07-15 18:06:29.540889] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:20:35.890 [2024-07-15 18:06:29.540956] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:20:35.890 [2024-07-15 18:06:29.540958] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@862 -- # return 0 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@728 -- # xtrace_disable 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@92 -- # adq_configure_nvmf_target 1 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@42 -- # rpc_cmd sock_get_default_impl 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@42 -- # jq -r .impl_name 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@42 -- # socket_impl=posix 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@43 -- # rpc_cmd sock_impl_set_options --enable-placement-id 1 --enable-zerocopy-send-server -i posix 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@44 -- # rpc_cmd framework_start_init 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@45 -- # rpc_cmd nvmf_create_transport -t tcp -o --io-unit-size 8192 --sock-priority 1 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:36.818 [2024-07-15 18:06:30.397856] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@46 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:36.818 Malloc1 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@47 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@48 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@49 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:36.818 [2024-07-15 18:06:30.445372] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@96 -- # perfpid=643623 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@97 -- # sleep 2 00:20:36.818 18:06:30 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@93 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 64 -o 4096 -w randread -t 10 -c 0xF0 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:20:36.818 EAL: No free 2048 kB hugepages reported on node 1 00:20:38.748 18:06:32 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@99 -- # rpc_cmd nvmf_get_stats 00:20:38.748 18:06:32 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:38.748 18:06:32 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:38.748 18:06:32 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:39.005 18:06:32 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@99 -- # nvmf_stats='{ 00:20:39.005 "tick_rate": 2300000000, 00:20:39.005 "poll_groups": [ 00:20:39.005 { 00:20:39.005 "name": "nvmf_tgt_poll_group_000", 00:20:39.005 "admin_qpairs": 1, 00:20:39.005 "io_qpairs": 1, 00:20:39.005 "current_admin_qpairs": 1, 00:20:39.005 "current_io_qpairs": 1, 00:20:39.005 "pending_bdev_io": 0, 00:20:39.005 "completed_nvme_io": 27757, 00:20:39.005 "transports": [ 00:20:39.005 { 00:20:39.005 "trtype": "TCP" 00:20:39.005 } 00:20:39.005 ] 00:20:39.005 }, 00:20:39.005 { 00:20:39.005 "name": "nvmf_tgt_poll_group_001", 00:20:39.005 "admin_qpairs": 0, 00:20:39.005 "io_qpairs": 3, 00:20:39.005 "current_admin_qpairs": 0, 00:20:39.005 "current_io_qpairs": 3, 00:20:39.005 "pending_bdev_io": 0, 00:20:39.005 "completed_nvme_io": 30195, 00:20:39.005 "transports": [ 00:20:39.005 { 00:20:39.005 "trtype": "TCP" 00:20:39.005 } 00:20:39.005 ] 00:20:39.005 }, 00:20:39.005 { 00:20:39.005 "name": "nvmf_tgt_poll_group_002", 00:20:39.005 "admin_qpairs": 0, 00:20:39.005 "io_qpairs": 0, 00:20:39.005 "current_admin_qpairs": 0, 00:20:39.005 "current_io_qpairs": 0, 00:20:39.005 "pending_bdev_io": 0, 00:20:39.005 "completed_nvme_io": 0, 00:20:39.005 "transports": [ 00:20:39.005 { 00:20:39.005 "trtype": "TCP" 00:20:39.005 } 00:20:39.005 ] 00:20:39.005 }, 00:20:39.005 { 00:20:39.005 "name": "nvmf_tgt_poll_group_003", 00:20:39.005 "admin_qpairs": 0, 00:20:39.005 "io_qpairs": 0, 00:20:39.005 "current_admin_qpairs": 0, 00:20:39.005 "current_io_qpairs": 0, 00:20:39.005 "pending_bdev_io": 0, 00:20:39.005 "completed_nvme_io": 0, 00:20:39.005 "transports": [ 00:20:39.005 { 00:20:39.005 "trtype": "TCP" 00:20:39.005 } 00:20:39.005 ] 00:20:39.005 } 00:20:39.005 ] 00:20:39.005 }' 00:20:39.005 18:06:32 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@100 -- # jq -r '.poll_groups[] | select(.current_io_qpairs == 0) | length' 00:20:39.005 18:06:32 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@100 -- # wc -l 00:20:39.005 18:06:32 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@100 -- # count=2 00:20:39.005 18:06:32 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@101 -- # [[ 2 -lt 2 ]] 00:20:39.005 18:06:32 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@106 -- # wait 643623 00:20:47.124 Initializing NVMe Controllers 00:20:47.124 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:20:47.124 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 4 00:20:47.124 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 5 00:20:47.124 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 6 00:20:47.124 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 7 00:20:47.124 Initialization complete. Launching workers. 00:20:47.124 ======================================================== 00:20:47.124 Latency(us) 00:20:47.124 Device Information : IOPS MiB/s Average min max 00:20:47.124 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 4: 14620.29 57.11 4377.49 1201.60 45432.52 00:20:47.124 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 5: 5173.66 20.21 12371.58 1634.25 57365.52 00:20:47.124 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 6: 5399.96 21.09 11870.03 1591.83 60031.43 00:20:47.124 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 7: 5390.76 21.06 11873.43 1573.64 57371.78 00:20:47.124 ======================================================== 00:20:47.124 Total : 30584.66 119.47 8373.84 1201.60 60031.43 00:20:47.124 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@107 -- # nvmftestfini 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@488 -- # nvmfcleanup 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@117 -- # sync 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@120 -- # set +e 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@121 -- # for i in {1..20} 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:20:47.124 rmmod nvme_tcp 00:20:47.124 rmmod nvme_fabrics 00:20:47.124 rmmod nvme_keyring 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@124 -- # set -e 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@125 -- # return 0 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@489 -- # '[' -n 643371 ']' 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@490 -- # killprocess 643371 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@948 -- # '[' -z 643371 ']' 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@952 -- # kill -0 643371 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@953 -- # uname 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 643371 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@966 -- # echo 'killing process with pid 643371' 00:20:47.124 killing process with pid 643371 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@967 -- # kill 643371 00:20:47.124 18:06:40 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@972 -- # wait 643371 00:20:47.382 18:06:40 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:20:47.382 18:06:40 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:20:47.382 18:06:40 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:20:47.382 18:06:40 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:20:47.382 18:06:40 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@278 -- # remove_spdk_ns 00:20:47.382 18:06:40 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:47.382 18:06:40 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:20:47.382 18:06:40 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:49.281 18:06:42 nvmf_tcp.nvmf_perf_adq -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:20:49.282 18:06:42 nvmf_tcp.nvmf_perf_adq -- target/perf_adq.sh@109 -- # trap - SIGINT SIGTERM EXIT 00:20:49.282 00:20:49.282 real 0m49.631s 00:20:49.282 user 2m48.935s 00:20:49.282 sys 0m9.390s 00:20:49.282 18:06:42 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@1124 -- # xtrace_disable 00:20:49.282 18:06:42 nvmf_tcp.nvmf_perf_adq -- common/autotest_common.sh@10 -- # set +x 00:20:49.282 ************************************ 00:20:49.282 END TEST nvmf_perf_adq 00:20:49.282 ************************************ 00:20:49.540 18:06:43 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:20:49.540 18:06:43 nvmf_tcp -- nvmf/nvmf.sh@83 -- # run_test nvmf_shutdown /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/shutdown.sh --transport=tcp 00:20:49.540 18:06:43 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:20:49.540 18:06:43 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:20:49.540 18:06:43 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:20:49.540 ************************************ 00:20:49.540 START TEST nvmf_shutdown 00:20:49.540 ************************************ 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/shutdown.sh --transport=tcp 00:20:49.540 * Looking for test storage... 00:20:49.540 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- target/shutdown.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@7 -- # uname -s 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- paths/export.sh@5 -- # export PATH 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@47 -- # : 0 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- nvmf/common.sh@51 -- # have_pci_nics=0 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- target/shutdown.sh@11 -- # MALLOC_BDEV_SIZE=64 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- target/shutdown.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- target/shutdown.sh@147 -- # run_test nvmf_shutdown_tc1 nvmf_shutdown_tc1 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- common/autotest_common.sh@1105 -- # xtrace_disable 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:20:49.540 ************************************ 00:20:49.540 START TEST nvmf_shutdown_tc1 00:20:49.540 ************************************ 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@1123 -- # nvmf_shutdown_tc1 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@74 -- # starttarget 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@15 -- # nvmftestinit 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@448 -- # prepare_net_devs 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@410 -- # local -g is_hw=no 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@412 -- # remove_spdk_ns 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@285 -- # xtrace_disable 00:20:49.540 18:06:43 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@291 -- # pci_devs=() 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@291 -- # local -a pci_devs 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@293 -- # pci_drivers=() 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@295 -- # net_devs=() 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@295 -- # local -ga net_devs 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@296 -- # e810=() 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@296 -- # local -ga e810 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@297 -- # x722=() 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@297 -- # local -ga x722 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@298 -- # mlx=() 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@298 -- # local -ga mlx 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:20:54.800 Found 0000:86:00.0 (0x8086 - 0x159b) 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:20:54.800 Found 0000:86:00.1 (0x8086 - 0x159b) 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:20:54.800 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@390 -- # [[ up == up ]] 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:20:54.801 Found net devices under 0000:86:00.0: cvl_0_0 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@390 -- # [[ up == up ]] 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:20:54.801 Found net devices under 0000:86:00.1: cvl_0_1 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@414 -- # is_hw=yes 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:20:54.801 18:06:47 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:20:54.801 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:20:54.801 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.198 ms 00:20:54.801 00:20:54.801 --- 10.0.0.2 ping statistics --- 00:20:54.801 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:54.801 rtt min/avg/max/mdev = 0.198/0.198/0.198/0.000 ms 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:20:54.801 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:20:54.801 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.202 ms 00:20:54.801 00:20:54.801 --- 10.0.0.1 ping statistics --- 00:20:54.801 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:20:54.801 rtt min/avg/max/mdev = 0.202/0.202/0.202/0.000 ms 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@422 -- # return 0 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@18 -- # nvmfappstart -m 0x1E 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@722 -- # xtrace_disable 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@481 -- # nvmfpid=648615 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@482 -- # waitforlisten 648615 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@829 -- # '[' -z 648615 ']' 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@834 -- # local max_retries=100 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:20:54.801 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@838 -- # xtrace_disable 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:20:54.801 18:06:48 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:20:54.801 [2024-07-15 18:06:48.241941] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:20:54.801 [2024-07-15 18:06:48.241988] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:20:54.801 EAL: No free 2048 kB hugepages reported on node 1 00:20:54.801 [2024-07-15 18:06:48.300062] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:20:54.801 [2024-07-15 18:06:48.380298] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:20:54.801 [2024-07-15 18:06:48.380336] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:20:54.801 [2024-07-15 18:06:48.380343] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:20:54.801 [2024-07-15 18:06:48.380350] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:20:54.801 [2024-07-15 18:06:48.380355] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:20:54.801 [2024-07-15 18:06:48.380454] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:20:54.801 [2024-07-15 18:06:48.380474] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:20:54.801 [2024-07-15 18:06:48.380603] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:20:54.801 [2024-07-15 18:06:48.380604] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 4 00:20:55.362 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:20:55.362 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@862 -- # return 0 00:20:55.362 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:20:55.362 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@728 -- # xtrace_disable 00:20:55.362 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@20 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:20:55.619 [2024-07-15 18:06:49.099268] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@22 -- # num_subsystems=({1..10}) 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@24 -- # timing_enter create_subsystems 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@722 -- # xtrace_disable 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@26 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # cat 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # cat 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # cat 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # cat 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # cat 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # cat 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # cat 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # cat 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # cat 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@28 -- # cat 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@35 -- # rpc_cmd 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:55.619 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:20:55.619 Malloc1 00:20:55.619 [2024-07-15 18:06:49.195016] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:20:55.619 Malloc2 00:20:55.619 Malloc3 00:20:55.619 Malloc4 00:20:55.619 Malloc5 00:20:55.875 Malloc6 00:20:55.875 Malloc7 00:20:55.875 Malloc8 00:20:55.875 Malloc9 00:20:55.875 Malloc10 00:20:55.875 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:55.875 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@36 -- # timing_exit create_subsystems 00:20:55.875 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@728 -- # xtrace_disable 00:20:55.875 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:20:56.134 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@78 -- # perfpid=648894 00:20:56.134 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@79 -- # waitforlisten 648894 /var/tmp/bdevperf.sock 00:20:56.134 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@829 -- # '[' -z 648894 ']' 00:20:56.134 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@77 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:20:56.134 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:20:56.134 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@532 -- # config=() 00:20:56.134 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/bdev_svc/bdev_svc -m 0x1 -i 1 -r /var/tmp/bdevperf.sock --json /dev/fd/63 00:20:56.134 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@532 -- # local subsystem config 00:20:56.134 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@834 -- # local max_retries=100 00:20:56.134 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:20:56.134 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:20:56.134 { 00:20:56.134 "params": { 00:20:56.134 "name": "Nvme$subsystem", 00:20:56.134 "trtype": "$TEST_TRANSPORT", 00:20:56.134 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:56.134 "adrfam": "ipv4", 00:20:56.134 "trsvcid": "$NVMF_PORT", 00:20:56.134 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:56.134 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:56.134 "hdgst": ${hdgst:-false}, 00:20:56.134 "ddgst": ${ddgst:-false} 00:20:56.134 }, 00:20:56.134 "method": "bdev_nvme_attach_controller" 00:20:56.134 } 00:20:56.134 EOF 00:20:56.134 )") 00:20:56.134 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:20:56.134 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@838 -- # xtrace_disable 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # cat 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:20:56.135 { 00:20:56.135 "params": { 00:20:56.135 "name": "Nvme$subsystem", 00:20:56.135 "trtype": "$TEST_TRANSPORT", 00:20:56.135 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:56.135 "adrfam": "ipv4", 00:20:56.135 "trsvcid": "$NVMF_PORT", 00:20:56.135 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:56.135 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:56.135 "hdgst": ${hdgst:-false}, 00:20:56.135 "ddgst": ${ddgst:-false} 00:20:56.135 }, 00:20:56.135 "method": "bdev_nvme_attach_controller" 00:20:56.135 } 00:20:56.135 EOF 00:20:56.135 )") 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # cat 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:20:56.135 { 00:20:56.135 "params": { 00:20:56.135 "name": "Nvme$subsystem", 00:20:56.135 "trtype": "$TEST_TRANSPORT", 00:20:56.135 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:56.135 "adrfam": "ipv4", 00:20:56.135 "trsvcid": "$NVMF_PORT", 00:20:56.135 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:56.135 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:56.135 "hdgst": ${hdgst:-false}, 00:20:56.135 "ddgst": ${ddgst:-false} 00:20:56.135 }, 00:20:56.135 "method": "bdev_nvme_attach_controller" 00:20:56.135 } 00:20:56.135 EOF 00:20:56.135 )") 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # cat 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:20:56.135 { 00:20:56.135 "params": { 00:20:56.135 "name": "Nvme$subsystem", 00:20:56.135 "trtype": "$TEST_TRANSPORT", 00:20:56.135 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:56.135 "adrfam": "ipv4", 00:20:56.135 "trsvcid": "$NVMF_PORT", 00:20:56.135 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:56.135 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:56.135 "hdgst": ${hdgst:-false}, 00:20:56.135 "ddgst": ${ddgst:-false} 00:20:56.135 }, 00:20:56.135 "method": "bdev_nvme_attach_controller" 00:20:56.135 } 00:20:56.135 EOF 00:20:56.135 )") 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # cat 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:20:56.135 { 00:20:56.135 "params": { 00:20:56.135 "name": "Nvme$subsystem", 00:20:56.135 "trtype": "$TEST_TRANSPORT", 00:20:56.135 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:56.135 "adrfam": "ipv4", 00:20:56.135 "trsvcid": "$NVMF_PORT", 00:20:56.135 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:56.135 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:56.135 "hdgst": ${hdgst:-false}, 00:20:56.135 "ddgst": ${ddgst:-false} 00:20:56.135 }, 00:20:56.135 "method": "bdev_nvme_attach_controller" 00:20:56.135 } 00:20:56.135 EOF 00:20:56.135 )") 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # cat 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:20:56.135 { 00:20:56.135 "params": { 00:20:56.135 "name": "Nvme$subsystem", 00:20:56.135 "trtype": "$TEST_TRANSPORT", 00:20:56.135 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:56.135 "adrfam": "ipv4", 00:20:56.135 "trsvcid": "$NVMF_PORT", 00:20:56.135 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:56.135 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:56.135 "hdgst": ${hdgst:-false}, 00:20:56.135 "ddgst": ${ddgst:-false} 00:20:56.135 }, 00:20:56.135 "method": "bdev_nvme_attach_controller" 00:20:56.135 } 00:20:56.135 EOF 00:20:56.135 )") 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # cat 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:20:56.135 { 00:20:56.135 "params": { 00:20:56.135 "name": "Nvme$subsystem", 00:20:56.135 "trtype": "$TEST_TRANSPORT", 00:20:56.135 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:56.135 "adrfam": "ipv4", 00:20:56.135 "trsvcid": "$NVMF_PORT", 00:20:56.135 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:56.135 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:56.135 "hdgst": ${hdgst:-false}, 00:20:56.135 "ddgst": ${ddgst:-false} 00:20:56.135 }, 00:20:56.135 "method": "bdev_nvme_attach_controller" 00:20:56.135 } 00:20:56.135 EOF 00:20:56.135 )") 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # cat 00:20:56.135 [2024-07-15 18:06:49.658806] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:20:56.135 [2024-07-15 18:06:49.658860] [ DPDK EAL parameters: bdev_svc -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk1 --proc-type=auto ] 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:20:56.135 { 00:20:56.135 "params": { 00:20:56.135 "name": "Nvme$subsystem", 00:20:56.135 "trtype": "$TEST_TRANSPORT", 00:20:56.135 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:56.135 "adrfam": "ipv4", 00:20:56.135 "trsvcid": "$NVMF_PORT", 00:20:56.135 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:56.135 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:56.135 "hdgst": ${hdgst:-false}, 00:20:56.135 "ddgst": ${ddgst:-false} 00:20:56.135 }, 00:20:56.135 "method": "bdev_nvme_attach_controller" 00:20:56.135 } 00:20:56.135 EOF 00:20:56.135 )") 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # cat 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:20:56.135 { 00:20:56.135 "params": { 00:20:56.135 "name": "Nvme$subsystem", 00:20:56.135 "trtype": "$TEST_TRANSPORT", 00:20:56.135 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:56.135 "adrfam": "ipv4", 00:20:56.135 "trsvcid": "$NVMF_PORT", 00:20:56.135 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:56.135 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:56.135 "hdgst": ${hdgst:-false}, 00:20:56.135 "ddgst": ${ddgst:-false} 00:20:56.135 }, 00:20:56.135 "method": "bdev_nvme_attach_controller" 00:20:56.135 } 00:20:56.135 EOF 00:20:56.135 )") 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # cat 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:20:56.135 { 00:20:56.135 "params": { 00:20:56.135 "name": "Nvme$subsystem", 00:20:56.135 "trtype": "$TEST_TRANSPORT", 00:20:56.135 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:56.135 "adrfam": "ipv4", 00:20:56.135 "trsvcid": "$NVMF_PORT", 00:20:56.135 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:56.135 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:56.135 "hdgst": ${hdgst:-false}, 00:20:56.135 "ddgst": ${ddgst:-false} 00:20:56.135 }, 00:20:56.135 "method": "bdev_nvme_attach_controller" 00:20:56.135 } 00:20:56.135 EOF 00:20:56.135 )") 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # cat 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@556 -- # jq . 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@557 -- # IFS=, 00:20:56.135 18:06:49 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:20:56.135 "params": { 00:20:56.135 "name": "Nvme1", 00:20:56.135 "trtype": "tcp", 00:20:56.135 "traddr": "10.0.0.2", 00:20:56.135 "adrfam": "ipv4", 00:20:56.135 "trsvcid": "4420", 00:20:56.135 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:20:56.135 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:20:56.135 "hdgst": false, 00:20:56.135 "ddgst": false 00:20:56.135 }, 00:20:56.135 "method": "bdev_nvme_attach_controller" 00:20:56.135 },{ 00:20:56.135 "params": { 00:20:56.135 "name": "Nvme2", 00:20:56.135 "trtype": "tcp", 00:20:56.135 "traddr": "10.0.0.2", 00:20:56.135 "adrfam": "ipv4", 00:20:56.135 "trsvcid": "4420", 00:20:56.135 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:20:56.135 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:20:56.135 "hdgst": false, 00:20:56.135 "ddgst": false 00:20:56.135 }, 00:20:56.135 "method": "bdev_nvme_attach_controller" 00:20:56.135 },{ 00:20:56.135 "params": { 00:20:56.135 "name": "Nvme3", 00:20:56.135 "trtype": "tcp", 00:20:56.135 "traddr": "10.0.0.2", 00:20:56.135 "adrfam": "ipv4", 00:20:56.135 "trsvcid": "4420", 00:20:56.135 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:20:56.135 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:20:56.135 "hdgst": false, 00:20:56.135 "ddgst": false 00:20:56.135 }, 00:20:56.135 "method": "bdev_nvme_attach_controller" 00:20:56.135 },{ 00:20:56.135 "params": { 00:20:56.135 "name": "Nvme4", 00:20:56.135 "trtype": "tcp", 00:20:56.135 "traddr": "10.0.0.2", 00:20:56.136 "adrfam": "ipv4", 00:20:56.136 "trsvcid": "4420", 00:20:56.136 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:20:56.136 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:20:56.136 "hdgst": false, 00:20:56.136 "ddgst": false 00:20:56.136 }, 00:20:56.136 "method": "bdev_nvme_attach_controller" 00:20:56.136 },{ 00:20:56.136 "params": { 00:20:56.136 "name": "Nvme5", 00:20:56.136 "trtype": "tcp", 00:20:56.136 "traddr": "10.0.0.2", 00:20:56.136 "adrfam": "ipv4", 00:20:56.136 "trsvcid": "4420", 00:20:56.136 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:20:56.136 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:20:56.136 "hdgst": false, 00:20:56.136 "ddgst": false 00:20:56.136 }, 00:20:56.136 "method": "bdev_nvme_attach_controller" 00:20:56.136 },{ 00:20:56.136 "params": { 00:20:56.136 "name": "Nvme6", 00:20:56.136 "trtype": "tcp", 00:20:56.136 "traddr": "10.0.0.2", 00:20:56.136 "adrfam": "ipv4", 00:20:56.136 "trsvcid": "4420", 00:20:56.136 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:20:56.136 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:20:56.136 "hdgst": false, 00:20:56.136 "ddgst": false 00:20:56.136 }, 00:20:56.136 "method": "bdev_nvme_attach_controller" 00:20:56.136 },{ 00:20:56.136 "params": { 00:20:56.136 "name": "Nvme7", 00:20:56.136 "trtype": "tcp", 00:20:56.136 "traddr": "10.0.0.2", 00:20:56.136 "adrfam": "ipv4", 00:20:56.136 "trsvcid": "4420", 00:20:56.136 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:20:56.136 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:20:56.136 "hdgst": false, 00:20:56.136 "ddgst": false 00:20:56.136 }, 00:20:56.136 "method": "bdev_nvme_attach_controller" 00:20:56.136 },{ 00:20:56.136 "params": { 00:20:56.136 "name": "Nvme8", 00:20:56.136 "trtype": "tcp", 00:20:56.136 "traddr": "10.0.0.2", 00:20:56.136 "adrfam": "ipv4", 00:20:56.136 "trsvcid": "4420", 00:20:56.136 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:20:56.136 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:20:56.136 "hdgst": false, 00:20:56.136 "ddgst": false 00:20:56.136 }, 00:20:56.136 "method": "bdev_nvme_attach_controller" 00:20:56.136 },{ 00:20:56.136 "params": { 00:20:56.136 "name": "Nvme9", 00:20:56.136 "trtype": "tcp", 00:20:56.136 "traddr": "10.0.0.2", 00:20:56.136 "adrfam": "ipv4", 00:20:56.136 "trsvcid": "4420", 00:20:56.136 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:20:56.136 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:20:56.136 "hdgst": false, 00:20:56.136 "ddgst": false 00:20:56.136 }, 00:20:56.136 "method": "bdev_nvme_attach_controller" 00:20:56.136 },{ 00:20:56.136 "params": { 00:20:56.136 "name": "Nvme10", 00:20:56.136 "trtype": "tcp", 00:20:56.136 "traddr": "10.0.0.2", 00:20:56.136 "adrfam": "ipv4", 00:20:56.136 "trsvcid": "4420", 00:20:56.136 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:20:56.136 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:20:56.136 "hdgst": false, 00:20:56.136 "ddgst": false 00:20:56.136 }, 00:20:56.136 "method": "bdev_nvme_attach_controller" 00:20:56.136 }' 00:20:56.136 EAL: No free 2048 kB hugepages reported on node 1 00:20:56.136 [2024-07-15 18:06:49.715710] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:56.136 [2024-07-15 18:06:49.790357] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:20:57.500 18:06:51 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:20:57.500 18:06:51 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@862 -- # return 0 00:20:57.500 18:06:51 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@80 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:20:57.500 18:06:51 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@559 -- # xtrace_disable 00:20:57.500 18:06:51 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:20:57.500 18:06:51 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:20:57.500 18:06:51 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@83 -- # kill -9 648894 00:20:57.500 18:06:51 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@84 -- # rm -f /var/run/spdk_bdev1 00:20:57.500 18:06:51 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@87 -- # sleep 1 00:20:58.426 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/shutdown.sh: line 73: 648894 Killed $rootdir/test/app/bdev_svc/bdev_svc -m 0x1 -i 1 -r /var/tmp/bdevperf.sock --json <(gen_nvmf_target_json "${num_subsystems[@]}") 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@88 -- # kill -0 648615 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@91 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 64 -o 65536 -w verify -t 1 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@91 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@532 -- # config=() 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@532 -- # local subsystem config 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:20:58.426 { 00:20:58.426 "params": { 00:20:58.426 "name": "Nvme$subsystem", 00:20:58.426 "trtype": "$TEST_TRANSPORT", 00:20:58.426 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:58.426 "adrfam": "ipv4", 00:20:58.426 "trsvcid": "$NVMF_PORT", 00:20:58.426 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:58.426 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:58.426 "hdgst": ${hdgst:-false}, 00:20:58.426 "ddgst": ${ddgst:-false} 00:20:58.426 }, 00:20:58.426 "method": "bdev_nvme_attach_controller" 00:20:58.426 } 00:20:58.426 EOF 00:20:58.426 )") 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # cat 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:20:58.426 { 00:20:58.426 "params": { 00:20:58.426 "name": "Nvme$subsystem", 00:20:58.426 "trtype": "$TEST_TRANSPORT", 00:20:58.426 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:58.426 "adrfam": "ipv4", 00:20:58.426 "trsvcid": "$NVMF_PORT", 00:20:58.426 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:58.426 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:58.426 "hdgst": ${hdgst:-false}, 00:20:58.426 "ddgst": ${ddgst:-false} 00:20:58.426 }, 00:20:58.426 "method": "bdev_nvme_attach_controller" 00:20:58.426 } 00:20:58.426 EOF 00:20:58.426 )") 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # cat 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:20:58.426 { 00:20:58.426 "params": { 00:20:58.426 "name": "Nvme$subsystem", 00:20:58.426 "trtype": "$TEST_TRANSPORT", 00:20:58.426 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:58.426 "adrfam": "ipv4", 00:20:58.426 "trsvcid": "$NVMF_PORT", 00:20:58.426 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:58.426 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:58.426 "hdgst": ${hdgst:-false}, 00:20:58.426 "ddgst": ${ddgst:-false} 00:20:58.426 }, 00:20:58.426 "method": "bdev_nvme_attach_controller" 00:20:58.426 } 00:20:58.426 EOF 00:20:58.426 )") 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # cat 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:20:58.426 { 00:20:58.426 "params": { 00:20:58.426 "name": "Nvme$subsystem", 00:20:58.426 "trtype": "$TEST_TRANSPORT", 00:20:58.426 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:58.426 "adrfam": "ipv4", 00:20:58.426 "trsvcid": "$NVMF_PORT", 00:20:58.426 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:58.426 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:58.426 "hdgst": ${hdgst:-false}, 00:20:58.426 "ddgst": ${ddgst:-false} 00:20:58.426 }, 00:20:58.426 "method": "bdev_nvme_attach_controller" 00:20:58.426 } 00:20:58.426 EOF 00:20:58.426 )") 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # cat 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:20:58.426 { 00:20:58.426 "params": { 00:20:58.426 "name": "Nvme$subsystem", 00:20:58.426 "trtype": "$TEST_TRANSPORT", 00:20:58.426 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:58.426 "adrfam": "ipv4", 00:20:58.426 "trsvcid": "$NVMF_PORT", 00:20:58.426 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:58.426 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:58.426 "hdgst": ${hdgst:-false}, 00:20:58.426 "ddgst": ${ddgst:-false} 00:20:58.426 }, 00:20:58.426 "method": "bdev_nvme_attach_controller" 00:20:58.426 } 00:20:58.426 EOF 00:20:58.426 )") 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # cat 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:20:58.426 { 00:20:58.426 "params": { 00:20:58.426 "name": "Nvme$subsystem", 00:20:58.426 "trtype": "$TEST_TRANSPORT", 00:20:58.426 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:58.426 "adrfam": "ipv4", 00:20:58.426 "trsvcid": "$NVMF_PORT", 00:20:58.426 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:58.426 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:58.426 "hdgst": ${hdgst:-false}, 00:20:58.426 "ddgst": ${ddgst:-false} 00:20:58.426 }, 00:20:58.426 "method": "bdev_nvme_attach_controller" 00:20:58.426 } 00:20:58.426 EOF 00:20:58.426 )") 00:20:58.426 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # cat 00:20:58.683 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:20:58.683 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:20:58.683 { 00:20:58.683 "params": { 00:20:58.683 "name": "Nvme$subsystem", 00:20:58.683 "trtype": "$TEST_TRANSPORT", 00:20:58.683 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:58.683 "adrfam": "ipv4", 00:20:58.683 "trsvcid": "$NVMF_PORT", 00:20:58.683 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:58.683 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:58.683 "hdgst": ${hdgst:-false}, 00:20:58.683 "ddgst": ${ddgst:-false} 00:20:58.683 }, 00:20:58.683 "method": "bdev_nvme_attach_controller" 00:20:58.683 } 00:20:58.683 EOF 00:20:58.683 )") 00:20:58.683 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # cat 00:20:58.683 [2024-07-15 18:06:52.158488] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:20:58.683 [2024-07-15 18:06:52.158537] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid649377 ] 00:20:58.683 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:20:58.683 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:20:58.683 { 00:20:58.683 "params": { 00:20:58.683 "name": "Nvme$subsystem", 00:20:58.683 "trtype": "$TEST_TRANSPORT", 00:20:58.683 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:58.683 "adrfam": "ipv4", 00:20:58.683 "trsvcid": "$NVMF_PORT", 00:20:58.683 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:58.683 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:58.683 "hdgst": ${hdgst:-false}, 00:20:58.683 "ddgst": ${ddgst:-false} 00:20:58.683 }, 00:20:58.683 "method": "bdev_nvme_attach_controller" 00:20:58.683 } 00:20:58.683 EOF 00:20:58.683 )") 00:20:58.683 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # cat 00:20:58.683 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:20:58.683 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:20:58.683 { 00:20:58.683 "params": { 00:20:58.683 "name": "Nvme$subsystem", 00:20:58.683 "trtype": "$TEST_TRANSPORT", 00:20:58.683 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:58.683 "adrfam": "ipv4", 00:20:58.683 "trsvcid": "$NVMF_PORT", 00:20:58.683 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:58.683 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:58.683 "hdgst": ${hdgst:-false}, 00:20:58.683 "ddgst": ${ddgst:-false} 00:20:58.683 }, 00:20:58.683 "method": "bdev_nvme_attach_controller" 00:20:58.683 } 00:20:58.683 EOF 00:20:58.683 )") 00:20:58.683 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # cat 00:20:58.683 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:20:58.683 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:20:58.683 { 00:20:58.683 "params": { 00:20:58.683 "name": "Nvme$subsystem", 00:20:58.683 "trtype": "$TEST_TRANSPORT", 00:20:58.683 "traddr": "$NVMF_FIRST_TARGET_IP", 00:20:58.683 "adrfam": "ipv4", 00:20:58.683 "trsvcid": "$NVMF_PORT", 00:20:58.683 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:20:58.683 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:20:58.683 "hdgst": ${hdgst:-false}, 00:20:58.683 "ddgst": ${ddgst:-false} 00:20:58.683 }, 00:20:58.683 "method": "bdev_nvme_attach_controller" 00:20:58.683 } 00:20:58.683 EOF 00:20:58.683 )") 00:20:58.683 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@554 -- # cat 00:20:58.683 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@556 -- # jq . 00:20:58.683 EAL: No free 2048 kB hugepages reported on node 1 00:20:58.683 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@557 -- # IFS=, 00:20:58.683 18:06:52 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:20:58.684 "params": { 00:20:58.684 "name": "Nvme1", 00:20:58.684 "trtype": "tcp", 00:20:58.684 "traddr": "10.0.0.2", 00:20:58.684 "adrfam": "ipv4", 00:20:58.684 "trsvcid": "4420", 00:20:58.684 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:20:58.684 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:20:58.684 "hdgst": false, 00:20:58.684 "ddgst": false 00:20:58.684 }, 00:20:58.684 "method": "bdev_nvme_attach_controller" 00:20:58.684 },{ 00:20:58.684 "params": { 00:20:58.684 "name": "Nvme2", 00:20:58.684 "trtype": "tcp", 00:20:58.684 "traddr": "10.0.0.2", 00:20:58.684 "adrfam": "ipv4", 00:20:58.684 "trsvcid": "4420", 00:20:58.684 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:20:58.684 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:20:58.684 "hdgst": false, 00:20:58.684 "ddgst": false 00:20:58.684 }, 00:20:58.684 "method": "bdev_nvme_attach_controller" 00:20:58.684 },{ 00:20:58.684 "params": { 00:20:58.684 "name": "Nvme3", 00:20:58.684 "trtype": "tcp", 00:20:58.684 "traddr": "10.0.0.2", 00:20:58.684 "adrfam": "ipv4", 00:20:58.684 "trsvcid": "4420", 00:20:58.684 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:20:58.684 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:20:58.684 "hdgst": false, 00:20:58.684 "ddgst": false 00:20:58.684 }, 00:20:58.684 "method": "bdev_nvme_attach_controller" 00:20:58.684 },{ 00:20:58.684 "params": { 00:20:58.684 "name": "Nvme4", 00:20:58.684 "trtype": "tcp", 00:20:58.684 "traddr": "10.0.0.2", 00:20:58.684 "adrfam": "ipv4", 00:20:58.684 "trsvcid": "4420", 00:20:58.684 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:20:58.684 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:20:58.684 "hdgst": false, 00:20:58.684 "ddgst": false 00:20:58.684 }, 00:20:58.684 "method": "bdev_nvme_attach_controller" 00:20:58.684 },{ 00:20:58.684 "params": { 00:20:58.684 "name": "Nvme5", 00:20:58.684 "trtype": "tcp", 00:20:58.684 "traddr": "10.0.0.2", 00:20:58.684 "adrfam": "ipv4", 00:20:58.684 "trsvcid": "4420", 00:20:58.684 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:20:58.684 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:20:58.684 "hdgst": false, 00:20:58.684 "ddgst": false 00:20:58.684 }, 00:20:58.684 "method": "bdev_nvme_attach_controller" 00:20:58.684 },{ 00:20:58.684 "params": { 00:20:58.684 "name": "Nvme6", 00:20:58.684 "trtype": "tcp", 00:20:58.684 "traddr": "10.0.0.2", 00:20:58.684 "adrfam": "ipv4", 00:20:58.684 "trsvcid": "4420", 00:20:58.684 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:20:58.684 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:20:58.684 "hdgst": false, 00:20:58.684 "ddgst": false 00:20:58.684 }, 00:20:58.684 "method": "bdev_nvme_attach_controller" 00:20:58.684 },{ 00:20:58.684 "params": { 00:20:58.684 "name": "Nvme7", 00:20:58.684 "trtype": "tcp", 00:20:58.684 "traddr": "10.0.0.2", 00:20:58.684 "adrfam": "ipv4", 00:20:58.684 "trsvcid": "4420", 00:20:58.684 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:20:58.684 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:20:58.684 "hdgst": false, 00:20:58.684 "ddgst": false 00:20:58.684 }, 00:20:58.684 "method": "bdev_nvme_attach_controller" 00:20:58.684 },{ 00:20:58.684 "params": { 00:20:58.684 "name": "Nvme8", 00:20:58.684 "trtype": "tcp", 00:20:58.684 "traddr": "10.0.0.2", 00:20:58.684 "adrfam": "ipv4", 00:20:58.684 "trsvcid": "4420", 00:20:58.684 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:20:58.684 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:20:58.684 "hdgst": false, 00:20:58.684 "ddgst": false 00:20:58.684 }, 00:20:58.684 "method": "bdev_nvme_attach_controller" 00:20:58.684 },{ 00:20:58.684 "params": { 00:20:58.684 "name": "Nvme9", 00:20:58.684 "trtype": "tcp", 00:20:58.684 "traddr": "10.0.0.2", 00:20:58.684 "adrfam": "ipv4", 00:20:58.684 "trsvcid": "4420", 00:20:58.684 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:20:58.684 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:20:58.684 "hdgst": false, 00:20:58.684 "ddgst": false 00:20:58.684 }, 00:20:58.684 "method": "bdev_nvme_attach_controller" 00:20:58.684 },{ 00:20:58.684 "params": { 00:20:58.684 "name": "Nvme10", 00:20:58.684 "trtype": "tcp", 00:20:58.684 "traddr": "10.0.0.2", 00:20:58.684 "adrfam": "ipv4", 00:20:58.684 "trsvcid": "4420", 00:20:58.684 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:20:58.684 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:20:58.684 "hdgst": false, 00:20:58.684 "ddgst": false 00:20:58.684 }, 00:20:58.684 "method": "bdev_nvme_attach_controller" 00:20:58.684 }' 00:20:58.684 [2024-07-15 18:06:52.214898] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:20:58.684 [2024-07-15 18:06:52.289563] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:21:00.572 Running I/O for 1 seconds... 00:21:01.502 00:21:01.502 Latency(us) 00:21:01.502 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:21:01.502 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:01.502 Verification LBA range: start 0x0 length 0x400 00:21:01.502 Nvme1n1 : 1.14 280.78 17.55 0.00 0.00 225774.55 16412.49 214274.23 00:21:01.502 Job: Nvme2n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:01.502 Verification LBA range: start 0x0 length 0x400 00:21:01.502 Nvme2n1 : 1.15 277.38 17.34 0.00 0.00 225476.79 15614.66 218833.25 00:21:01.502 Job: Nvme3n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:01.502 Verification LBA range: start 0x0 length 0x400 00:21:01.502 Nvme3n1 : 1.15 278.74 17.42 0.00 0.00 221214.81 20515.62 206979.78 00:21:01.502 Job: Nvme4n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:01.502 Verification LBA range: start 0x0 length 0x400 00:21:01.502 Nvme4n1 : 1.13 285.60 17.85 0.00 0.00 205474.66 6696.07 206067.98 00:21:01.502 Job: Nvme5n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:01.502 Verification LBA range: start 0x0 length 0x400 00:21:01.502 Nvme5n1 : 1.16 274.80 17.17 0.00 0.00 218283.32 18236.10 218833.25 00:21:01.502 Job: Nvme6n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:01.502 Verification LBA range: start 0x0 length 0x400 00:21:01.502 Nvme6n1 : 1.17 273.27 17.08 0.00 0.00 216387.41 18008.15 219745.06 00:21:01.502 Job: Nvme7n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:01.502 Verification LBA range: start 0x0 length 0x400 00:21:01.502 Nvme7n1 : 1.16 275.80 17.24 0.00 0.00 211082.82 13506.11 221568.67 00:21:01.502 Job: Nvme8n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:01.502 Verification LBA range: start 0x0 length 0x400 00:21:01.502 Nvme8n1 : 1.17 278.63 17.41 0.00 0.00 205344.63 4359.57 218833.25 00:21:01.502 Job: Nvme9n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:01.502 Verification LBA range: start 0x0 length 0x400 00:21:01.502 Nvme9n1 : 1.21 263.71 16.48 0.00 0.00 207738.79 29633.67 224304.08 00:21:01.502 Job: Nvme10n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:01.502 Verification LBA range: start 0x0 length 0x400 00:21:01.502 Nvme10n1 : 1.18 271.86 16.99 0.00 0.00 205078.88 17666.23 240716.58 00:21:01.502 =================================================================================================================== 00:21:01.502 Total : 2760.57 172.54 0.00 0.00 214161.03 4359.57 240716.58 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@94 -- # stoptarget 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@41 -- # rm -f ./local-job0-0-verify.state 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@42 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- target/shutdown.sh@45 -- # nvmftestfini 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@488 -- # nvmfcleanup 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@117 -- # sync 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@120 -- # set +e 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@121 -- # for i in {1..20} 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:21:01.794 rmmod nvme_tcp 00:21:01.794 rmmod nvme_fabrics 00:21:01.794 rmmod nvme_keyring 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@124 -- # set -e 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@125 -- # return 0 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@489 -- # '[' -n 648615 ']' 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@490 -- # killprocess 648615 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@948 -- # '[' -z 648615 ']' 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@952 -- # kill -0 648615 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@953 -- # uname 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 648615 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@966 -- # echo 'killing process with pid 648615' 00:21:01.794 killing process with pid 648615 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@967 -- # kill 648615 00:21:01.794 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@972 -- # wait 648615 00:21:02.065 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:21:02.065 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:21:02.065 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:21:02.065 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:21:02.065 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:21:02.065 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:02.065 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:02.065 18:06:55 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:21:04.593 00:21:04.593 real 0m14.629s 00:21:04.593 user 0m34.908s 00:21:04.593 sys 0m5.062s 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@1124 -- # xtrace_disable 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc1 -- common/autotest_common.sh@10 -- # set +x 00:21:04.593 ************************************ 00:21:04.593 END TEST nvmf_shutdown_tc1 00:21:04.593 ************************************ 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown -- common/autotest_common.sh@1142 -- # return 0 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown -- target/shutdown.sh@148 -- # run_test nvmf_shutdown_tc2 nvmf_shutdown_tc2 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown -- common/autotest_common.sh@1105 -- # xtrace_disable 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:21:04.593 ************************************ 00:21:04.593 START TEST nvmf_shutdown_tc2 00:21:04.593 ************************************ 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@1123 -- # nvmf_shutdown_tc2 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@99 -- # starttarget 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@15 -- # nvmftestinit 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@448 -- # prepare_net_devs 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@410 -- # local -g is_hw=no 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@412 -- # remove_spdk_ns 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@285 -- # xtrace_disable 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@291 -- # pci_devs=() 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@291 -- # local -a pci_devs 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@293 -- # pci_drivers=() 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@295 -- # net_devs=() 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@295 -- # local -ga net_devs 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@296 -- # e810=() 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@296 -- # local -ga e810 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@297 -- # x722=() 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@297 -- # local -ga x722 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@298 -- # mlx=() 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@298 -- # local -ga mlx 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:21:04.593 Found 0000:86:00.0 (0x8086 - 0x159b) 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:21:04.593 Found 0000:86:00.1 (0x8086 - 0x159b) 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@390 -- # [[ up == up ]] 00:21:04.593 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:21:04.594 Found net devices under 0000:86:00.0: cvl_0_0 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@390 -- # [[ up == up ]] 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:21:04.594 Found net devices under 0000:86:00.1: cvl_0_1 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@414 -- # is_hw=yes 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:21:04.594 18:06:57 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:21:04.594 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:21:04.594 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.186 ms 00:21:04.594 00:21:04.594 --- 10.0.0.2 ping statistics --- 00:21:04.594 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:04.594 rtt min/avg/max/mdev = 0.186/0.186/0.186/0.000 ms 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:21:04.594 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:21:04.594 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.238 ms 00:21:04.594 00:21:04.594 --- 10.0.0.1 ping statistics --- 00:21:04.594 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:04.594 rtt min/avg/max/mdev = 0.238/0.238/0.238/0.000 ms 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@422 -- # return 0 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@18 -- # nvmfappstart -m 0x1E 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@722 -- # xtrace_disable 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@481 -- # nvmfpid=650497 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@482 -- # waitforlisten 650497 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@829 -- # '[' -z 650497 ']' 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@834 -- # local max_retries=100 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:04.594 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@838 -- # xtrace_disable 00:21:04.594 18:06:58 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:21:04.594 [2024-07-15 18:06:58.276118] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:21:04.594 [2024-07-15 18:06:58.276160] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:04.594 EAL: No free 2048 kB hugepages reported on node 1 00:21:04.851 [2024-07-15 18:06:58.334694] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:21:04.851 [2024-07-15 18:06:58.410044] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:04.851 [2024-07-15 18:06:58.410084] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:04.851 [2024-07-15 18:06:58.410091] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:04.851 [2024-07-15 18:06:58.410096] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:04.851 [2024-07-15 18:06:58.410101] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:04.851 [2024-07-15 18:06:58.410217] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:21:04.851 [2024-07-15 18:06:58.410304] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:21:04.851 [2024-07-15 18:06:58.410341] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:21:04.851 [2024-07-15 18:06:58.410342] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 4 00:21:05.413 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:21:05.413 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@862 -- # return 0 00:21:05.413 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:21:05.413 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@728 -- # xtrace_disable 00:21:05.413 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:21:05.413 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:21:05.413 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@20 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:21:05.413 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:05.413 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:21:05.413 [2024-07-15 18:06:59.117102] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:21:05.413 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:05.413 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@22 -- # num_subsystems=({1..10}) 00:21:05.413 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@24 -- # timing_enter create_subsystems 00:21:05.413 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@722 -- # xtrace_disable 00:21:05.413 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:21:05.413 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@26 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:21:05.413 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:05.413 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # cat 00:21:05.413 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:05.413 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # cat 00:21:05.669 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:05.669 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # cat 00:21:05.669 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:05.669 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # cat 00:21:05.669 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:05.669 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # cat 00:21:05.669 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:05.669 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # cat 00:21:05.669 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:05.670 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # cat 00:21:05.670 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:05.670 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # cat 00:21:05.670 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:05.670 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # cat 00:21:05.670 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:05.670 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@28 -- # cat 00:21:05.670 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@35 -- # rpc_cmd 00:21:05.670 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:05.670 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:21:05.670 Malloc1 00:21:05.670 [2024-07-15 18:06:59.212715] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:21:05.670 Malloc2 00:21:05.670 Malloc3 00:21:05.670 Malloc4 00:21:05.670 Malloc5 00:21:05.926 Malloc6 00:21:05.926 Malloc7 00:21:05.926 Malloc8 00:21:05.926 Malloc9 00:21:05.926 Malloc10 00:21:05.926 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:05.926 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@36 -- # timing_exit create_subsystems 00:21:05.926 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@728 -- # xtrace_disable 00:21:05.926 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:21:05.926 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@103 -- # perfpid=650781 00:21:05.926 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@104 -- # waitforlisten 650781 /var/tmp/bdevperf.sock 00:21:05.926 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@829 -- # '[' -z 650781 ']' 00:21:05.926 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:21:05.926 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@102 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:21:05.926 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@834 -- # local max_retries=100 00:21:05.926 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@102 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:21:05.926 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:21:05.926 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:21:05.926 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@532 -- # config=() 00:21:05.926 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@838 -- # xtrace_disable 00:21:05.926 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@532 -- # local subsystem config 00:21:05.926 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:21:05.926 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:21:05.926 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:21:05.926 { 00:21:05.926 "params": { 00:21:05.926 "name": "Nvme$subsystem", 00:21:05.926 "trtype": "$TEST_TRANSPORT", 00:21:05.926 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:05.926 "adrfam": "ipv4", 00:21:05.926 "trsvcid": "$NVMF_PORT", 00:21:05.926 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:05.926 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:05.926 "hdgst": ${hdgst:-false}, 00:21:05.926 "ddgst": ${ddgst:-false} 00:21:05.926 }, 00:21:05.926 "method": "bdev_nvme_attach_controller" 00:21:05.926 } 00:21:05.926 EOF 00:21:05.926 )") 00:21:05.926 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@554 -- # cat 00:21:06.183 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:21:06.183 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:21:06.183 { 00:21:06.183 "params": { 00:21:06.183 "name": "Nvme$subsystem", 00:21:06.183 "trtype": "$TEST_TRANSPORT", 00:21:06.183 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:06.183 "adrfam": "ipv4", 00:21:06.183 "trsvcid": "$NVMF_PORT", 00:21:06.183 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:06.183 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:06.183 "hdgst": ${hdgst:-false}, 00:21:06.183 "ddgst": ${ddgst:-false} 00:21:06.183 }, 00:21:06.183 "method": "bdev_nvme_attach_controller" 00:21:06.183 } 00:21:06.183 EOF 00:21:06.183 )") 00:21:06.183 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@554 -- # cat 00:21:06.183 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:21:06.183 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:21:06.183 { 00:21:06.183 "params": { 00:21:06.183 "name": "Nvme$subsystem", 00:21:06.183 "trtype": "$TEST_TRANSPORT", 00:21:06.183 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:06.183 "adrfam": "ipv4", 00:21:06.183 "trsvcid": "$NVMF_PORT", 00:21:06.183 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:06.183 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:06.183 "hdgst": ${hdgst:-false}, 00:21:06.183 "ddgst": ${ddgst:-false} 00:21:06.183 }, 00:21:06.183 "method": "bdev_nvme_attach_controller" 00:21:06.183 } 00:21:06.184 EOF 00:21:06.184 )") 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@554 -- # cat 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:21:06.184 { 00:21:06.184 "params": { 00:21:06.184 "name": "Nvme$subsystem", 00:21:06.184 "trtype": "$TEST_TRANSPORT", 00:21:06.184 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:06.184 "adrfam": "ipv4", 00:21:06.184 "trsvcid": "$NVMF_PORT", 00:21:06.184 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:06.184 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:06.184 "hdgst": ${hdgst:-false}, 00:21:06.184 "ddgst": ${ddgst:-false} 00:21:06.184 }, 00:21:06.184 "method": "bdev_nvme_attach_controller" 00:21:06.184 } 00:21:06.184 EOF 00:21:06.184 )") 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@554 -- # cat 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:21:06.184 { 00:21:06.184 "params": { 00:21:06.184 "name": "Nvme$subsystem", 00:21:06.184 "trtype": "$TEST_TRANSPORT", 00:21:06.184 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:06.184 "adrfam": "ipv4", 00:21:06.184 "trsvcid": "$NVMF_PORT", 00:21:06.184 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:06.184 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:06.184 "hdgst": ${hdgst:-false}, 00:21:06.184 "ddgst": ${ddgst:-false} 00:21:06.184 }, 00:21:06.184 "method": "bdev_nvme_attach_controller" 00:21:06.184 } 00:21:06.184 EOF 00:21:06.184 )") 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@554 -- # cat 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:21:06.184 { 00:21:06.184 "params": { 00:21:06.184 "name": "Nvme$subsystem", 00:21:06.184 "trtype": "$TEST_TRANSPORT", 00:21:06.184 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:06.184 "adrfam": "ipv4", 00:21:06.184 "trsvcid": "$NVMF_PORT", 00:21:06.184 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:06.184 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:06.184 "hdgst": ${hdgst:-false}, 00:21:06.184 "ddgst": ${ddgst:-false} 00:21:06.184 }, 00:21:06.184 "method": "bdev_nvme_attach_controller" 00:21:06.184 } 00:21:06.184 EOF 00:21:06.184 )") 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@554 -- # cat 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:21:06.184 { 00:21:06.184 "params": { 00:21:06.184 "name": "Nvme$subsystem", 00:21:06.184 "trtype": "$TEST_TRANSPORT", 00:21:06.184 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:06.184 "adrfam": "ipv4", 00:21:06.184 "trsvcid": "$NVMF_PORT", 00:21:06.184 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:06.184 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:06.184 "hdgst": ${hdgst:-false}, 00:21:06.184 "ddgst": ${ddgst:-false} 00:21:06.184 }, 00:21:06.184 "method": "bdev_nvme_attach_controller" 00:21:06.184 } 00:21:06.184 EOF 00:21:06.184 )") 00:21:06.184 [2024-07-15 18:06:59.687054] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:21:06.184 [2024-07-15 18:06:59.687102] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid650781 ] 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@554 -- # cat 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:21:06.184 { 00:21:06.184 "params": { 00:21:06.184 "name": "Nvme$subsystem", 00:21:06.184 "trtype": "$TEST_TRANSPORT", 00:21:06.184 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:06.184 "adrfam": "ipv4", 00:21:06.184 "trsvcid": "$NVMF_PORT", 00:21:06.184 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:06.184 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:06.184 "hdgst": ${hdgst:-false}, 00:21:06.184 "ddgst": ${ddgst:-false} 00:21:06.184 }, 00:21:06.184 "method": "bdev_nvme_attach_controller" 00:21:06.184 } 00:21:06.184 EOF 00:21:06.184 )") 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@554 -- # cat 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:21:06.184 { 00:21:06.184 "params": { 00:21:06.184 "name": "Nvme$subsystem", 00:21:06.184 "trtype": "$TEST_TRANSPORT", 00:21:06.184 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:06.184 "adrfam": "ipv4", 00:21:06.184 "trsvcid": "$NVMF_PORT", 00:21:06.184 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:06.184 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:06.184 "hdgst": ${hdgst:-false}, 00:21:06.184 "ddgst": ${ddgst:-false} 00:21:06.184 }, 00:21:06.184 "method": "bdev_nvme_attach_controller" 00:21:06.184 } 00:21:06.184 EOF 00:21:06.184 )") 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@554 -- # cat 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:21:06.184 { 00:21:06.184 "params": { 00:21:06.184 "name": "Nvme$subsystem", 00:21:06.184 "trtype": "$TEST_TRANSPORT", 00:21:06.184 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:06.184 "adrfam": "ipv4", 00:21:06.184 "trsvcid": "$NVMF_PORT", 00:21:06.184 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:06.184 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:06.184 "hdgst": ${hdgst:-false}, 00:21:06.184 "ddgst": ${ddgst:-false} 00:21:06.184 }, 00:21:06.184 "method": "bdev_nvme_attach_controller" 00:21:06.184 } 00:21:06.184 EOF 00:21:06.184 )") 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@554 -- # cat 00:21:06.184 EAL: No free 2048 kB hugepages reported on node 1 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@556 -- # jq . 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@557 -- # IFS=, 00:21:06.184 18:06:59 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:21:06.184 "params": { 00:21:06.184 "name": "Nvme1", 00:21:06.184 "trtype": "tcp", 00:21:06.184 "traddr": "10.0.0.2", 00:21:06.184 "adrfam": "ipv4", 00:21:06.184 "trsvcid": "4420", 00:21:06.184 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:21:06.184 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:21:06.184 "hdgst": false, 00:21:06.184 "ddgst": false 00:21:06.184 }, 00:21:06.184 "method": "bdev_nvme_attach_controller" 00:21:06.184 },{ 00:21:06.184 "params": { 00:21:06.184 "name": "Nvme2", 00:21:06.184 "trtype": "tcp", 00:21:06.184 "traddr": "10.0.0.2", 00:21:06.184 "adrfam": "ipv4", 00:21:06.184 "trsvcid": "4420", 00:21:06.184 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:21:06.184 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:21:06.184 "hdgst": false, 00:21:06.184 "ddgst": false 00:21:06.184 }, 00:21:06.184 "method": "bdev_nvme_attach_controller" 00:21:06.184 },{ 00:21:06.184 "params": { 00:21:06.184 "name": "Nvme3", 00:21:06.184 "trtype": "tcp", 00:21:06.184 "traddr": "10.0.0.2", 00:21:06.184 "adrfam": "ipv4", 00:21:06.184 "trsvcid": "4420", 00:21:06.184 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:21:06.184 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:21:06.184 "hdgst": false, 00:21:06.184 "ddgst": false 00:21:06.184 }, 00:21:06.184 "method": "bdev_nvme_attach_controller" 00:21:06.184 },{ 00:21:06.184 "params": { 00:21:06.184 "name": "Nvme4", 00:21:06.184 "trtype": "tcp", 00:21:06.184 "traddr": "10.0.0.2", 00:21:06.184 "adrfam": "ipv4", 00:21:06.184 "trsvcid": "4420", 00:21:06.184 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:21:06.184 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:21:06.184 "hdgst": false, 00:21:06.184 "ddgst": false 00:21:06.184 }, 00:21:06.184 "method": "bdev_nvme_attach_controller" 00:21:06.184 },{ 00:21:06.184 "params": { 00:21:06.184 "name": "Nvme5", 00:21:06.184 "trtype": "tcp", 00:21:06.184 "traddr": "10.0.0.2", 00:21:06.184 "adrfam": "ipv4", 00:21:06.184 "trsvcid": "4420", 00:21:06.184 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:21:06.184 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:21:06.184 "hdgst": false, 00:21:06.184 "ddgst": false 00:21:06.184 }, 00:21:06.185 "method": "bdev_nvme_attach_controller" 00:21:06.185 },{ 00:21:06.185 "params": { 00:21:06.185 "name": "Nvme6", 00:21:06.185 "trtype": "tcp", 00:21:06.185 "traddr": "10.0.0.2", 00:21:06.185 "adrfam": "ipv4", 00:21:06.185 "trsvcid": "4420", 00:21:06.185 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:21:06.185 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:21:06.185 "hdgst": false, 00:21:06.185 "ddgst": false 00:21:06.185 }, 00:21:06.185 "method": "bdev_nvme_attach_controller" 00:21:06.185 },{ 00:21:06.185 "params": { 00:21:06.185 "name": "Nvme7", 00:21:06.185 "trtype": "tcp", 00:21:06.185 "traddr": "10.0.0.2", 00:21:06.185 "adrfam": "ipv4", 00:21:06.185 "trsvcid": "4420", 00:21:06.185 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:21:06.185 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:21:06.185 "hdgst": false, 00:21:06.185 "ddgst": false 00:21:06.185 }, 00:21:06.185 "method": "bdev_nvme_attach_controller" 00:21:06.185 },{ 00:21:06.185 "params": { 00:21:06.185 "name": "Nvme8", 00:21:06.185 "trtype": "tcp", 00:21:06.185 "traddr": "10.0.0.2", 00:21:06.185 "adrfam": "ipv4", 00:21:06.185 "trsvcid": "4420", 00:21:06.185 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:21:06.185 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:21:06.185 "hdgst": false, 00:21:06.185 "ddgst": false 00:21:06.185 }, 00:21:06.185 "method": "bdev_nvme_attach_controller" 00:21:06.185 },{ 00:21:06.185 "params": { 00:21:06.185 "name": "Nvme9", 00:21:06.185 "trtype": "tcp", 00:21:06.185 "traddr": "10.0.0.2", 00:21:06.185 "adrfam": "ipv4", 00:21:06.185 "trsvcid": "4420", 00:21:06.185 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:21:06.185 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:21:06.185 "hdgst": false, 00:21:06.185 "ddgst": false 00:21:06.185 }, 00:21:06.185 "method": "bdev_nvme_attach_controller" 00:21:06.185 },{ 00:21:06.185 "params": { 00:21:06.185 "name": "Nvme10", 00:21:06.185 "trtype": "tcp", 00:21:06.185 "traddr": "10.0.0.2", 00:21:06.185 "adrfam": "ipv4", 00:21:06.185 "trsvcid": "4420", 00:21:06.185 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:21:06.185 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:21:06.185 "hdgst": false, 00:21:06.185 "ddgst": false 00:21:06.185 }, 00:21:06.185 "method": "bdev_nvme_attach_controller" 00:21:06.185 }' 00:21:06.185 [2024-07-15 18:06:59.742842] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:06.185 [2024-07-15 18:06:59.816717] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:21:08.077 Running I/O for 10 seconds... 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@862 -- # return 0 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@105 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@107 -- # waitforio /var/tmp/bdevperf.sock Nvme1n1 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@50 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@54 -- # '[' -z Nvme1n1 ']' 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@57 -- # local ret=1 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@58 -- # local i 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@59 -- # (( i = 10 )) 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@59 -- # (( i != 0 )) 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # jq -r '.bdevs[0].num_read_ops' 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@60 -- # read_io_count=131 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@63 -- # '[' 131 -ge 100 ']' 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@64 -- # ret=0 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@65 -- # break 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@69 -- # return 0 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@110 -- # killprocess 650781 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@948 -- # '[' -z 650781 ']' 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@952 -- # kill -0 650781 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@953 -- # uname 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 650781 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@966 -- # echo 'killing process with pid 650781' 00:21:08.641 killing process with pid 650781 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@967 -- # kill 650781 00:21:08.641 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@972 -- # wait 650781 00:21:08.898 Received shutdown signal, test time was about 0.810666 seconds 00:21:08.898 00:21:08.898 Latency(us) 00:21:08.898 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:21:08.898 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:08.898 Verification LBA range: start 0x0 length 0x400 00:21:08.898 Nvme1n1 : 0.81 317.27 19.83 0.00 0.00 199324.72 16412.49 215186.03 00:21:08.898 Job: Nvme2n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:08.898 Verification LBA range: start 0x0 length 0x400 00:21:08.898 Nvme2n1 : 0.79 243.31 15.21 0.00 0.00 254596.38 26100.42 214274.23 00:21:08.898 Job: Nvme3n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:08.898 Verification LBA range: start 0x0 length 0x400 00:21:08.898 Nvme3n1 : 0.78 261.56 16.35 0.00 0.00 228932.95 8206.25 208803.39 00:21:08.898 Job: Nvme4n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:08.898 Verification LBA range: start 0x0 length 0x400 00:21:08.898 Nvme4n1 : 0.80 319.77 19.99 0.00 0.00 185921.56 13734.07 217921.45 00:21:08.898 Job: Nvme5n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:08.898 Verification LBA range: start 0x0 length 0x400 00:21:08.898 Nvme5n1 : 0.79 244.32 15.27 0.00 0.00 237901.32 16868.40 217009.64 00:21:08.898 Job: Nvme6n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:08.898 Verification LBA range: start 0x0 length 0x400 00:21:08.898 Nvme6n1 : 0.81 316.05 19.75 0.00 0.00 180382.94 15728.64 200597.15 00:21:08.898 Job: Nvme7n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:08.898 Verification LBA range: start 0x0 length 0x400 00:21:08.898 Nvme7n1 : 0.80 318.34 19.90 0.00 0.00 175001.38 15842.62 212450.62 00:21:08.898 Job: Nvme8n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:08.898 Verification LBA range: start 0x0 length 0x400 00:21:08.898 Nvme8n1 : 0.78 247.63 15.48 0.00 0.00 218632.31 14189.97 208803.39 00:21:08.898 Job: Nvme9n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:08.898 Verification LBA range: start 0x0 length 0x400 00:21:08.898 Nvme9n1 : 0.79 241.64 15.10 0.00 0.00 219791.36 21769.35 225215.89 00:21:08.898 Job: Nvme10n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:08.898 Verification LBA range: start 0x0 length 0x400 00:21:08.898 Nvme10n1 : 0.80 240.36 15.02 0.00 0.00 215837.24 17780.20 238892.97 00:21:08.898 =================================================================================================================== 00:21:08.898 Total : 2750.26 171.89 0.00 0.00 208629.53 8206.25 238892.97 00:21:08.898 18:07:02 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@113 -- # sleep 1 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@114 -- # kill -0 650497 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@116 -- # stoptarget 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@41 -- # rm -f ./local-job0-0-verify.state 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@42 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- target/shutdown.sh@45 -- # nvmftestfini 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@488 -- # nvmfcleanup 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@117 -- # sync 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@120 -- # set +e 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@121 -- # for i in {1..20} 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:21:10.268 rmmod nvme_tcp 00:21:10.268 rmmod nvme_fabrics 00:21:10.268 rmmod nvme_keyring 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@124 -- # set -e 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@125 -- # return 0 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@489 -- # '[' -n 650497 ']' 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@490 -- # killprocess 650497 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@948 -- # '[' -z 650497 ']' 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@952 -- # kill -0 650497 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@953 -- # uname 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 650497 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@966 -- # echo 'killing process with pid 650497' 00:21:10.268 killing process with pid 650497 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@967 -- # kill 650497 00:21:10.268 18:07:03 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@972 -- # wait 650497 00:21:10.527 18:07:04 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:21:10.527 18:07:04 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:21:10.527 18:07:04 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:21:10.527 18:07:04 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:21:10.527 18:07:04 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:21:10.527 18:07:04 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:10.527 18:07:04 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:10.527 18:07:04 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:12.426 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:21:12.426 00:21:12.426 real 0m8.240s 00:21:12.426 user 0m25.568s 00:21:12.426 sys 0m1.286s 00:21:12.426 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@1124 -- # xtrace_disable 00:21:12.426 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc2 -- common/autotest_common.sh@10 -- # set +x 00:21:12.426 ************************************ 00:21:12.426 END TEST nvmf_shutdown_tc2 00:21:12.426 ************************************ 00:21:12.684 18:07:06 nvmf_tcp.nvmf_shutdown -- common/autotest_common.sh@1142 -- # return 0 00:21:12.684 18:07:06 nvmf_tcp.nvmf_shutdown -- target/shutdown.sh@149 -- # run_test nvmf_shutdown_tc3 nvmf_shutdown_tc3 00:21:12.684 18:07:06 nvmf_tcp.nvmf_shutdown -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:21:12.684 18:07:06 nvmf_tcp.nvmf_shutdown -- common/autotest_common.sh@1105 -- # xtrace_disable 00:21:12.684 18:07:06 nvmf_tcp.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:21:12.684 ************************************ 00:21:12.684 START TEST nvmf_shutdown_tc3 00:21:12.684 ************************************ 00:21:12.684 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@1123 -- # nvmf_shutdown_tc3 00:21:12.684 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@121 -- # starttarget 00:21:12.684 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@15 -- # nvmftestinit 00:21:12.684 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@448 -- # prepare_net_devs 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@410 -- # local -g is_hw=no 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@412 -- # remove_spdk_ns 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@285 -- # xtrace_disable 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@291 -- # pci_devs=() 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@291 -- # local -a pci_devs 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@292 -- # pci_net_devs=() 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@293 -- # pci_drivers=() 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@293 -- # local -A pci_drivers 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@295 -- # net_devs=() 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@295 -- # local -ga net_devs 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@296 -- # e810=() 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@296 -- # local -ga e810 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@297 -- # x722=() 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@297 -- # local -ga x722 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@298 -- # mlx=() 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@298 -- # local -ga mlx 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:21:12.685 Found 0000:86:00.0 (0x8086 - 0x159b) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:21:12.685 Found 0000:86:00.1 (0x8086 - 0x159b) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@390 -- # [[ up == up ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:21:12.685 Found net devices under 0000:86:00.0: cvl_0_0 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@390 -- # [[ up == up ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:21:12.685 Found net devices under 0000:86:00.1: cvl_0_1 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@414 -- # is_hw=yes 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:21:12.685 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:21:12.943 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:21:12.943 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.176 ms 00:21:12.943 00:21:12.943 --- 10.0.0.2 ping statistics --- 00:21:12.943 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:12.943 rtt min/avg/max/mdev = 0.176/0.176/0.176/0.000 ms 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:21:12.943 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:21:12.943 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.235 ms 00:21:12.943 00:21:12.943 --- 10.0.0.1 ping statistics --- 00:21:12.943 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:12.943 rtt min/avg/max/mdev = 0.235/0.235/0.235/0.000 ms 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@422 -- # return 0 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@18 -- # nvmfappstart -m 0x1E 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@722 -- # xtrace_disable 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@481 -- # nvmfpid=651945 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@482 -- # waitforlisten 651945 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1E 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@829 -- # '[' -z 651945 ']' 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@834 -- # local max_retries=100 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:12.943 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@838 -- # xtrace_disable 00:21:12.943 18:07:06 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:21:12.943 [2024-07-15 18:07:06.560272] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:21:12.944 [2024-07-15 18:07:06.560332] [ DPDK EAL parameters: nvmf -c 0x1E --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:12.944 EAL: No free 2048 kB hugepages reported on node 1 00:21:12.944 [2024-07-15 18:07:06.617773] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:21:13.201 [2024-07-15 18:07:06.698297] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:13.201 [2024-07-15 18:07:06.698329] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:13.201 [2024-07-15 18:07:06.698336] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:13.201 [2024-07-15 18:07:06.698343] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:13.201 [2024-07-15 18:07:06.698348] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:13.201 [2024-07-15 18:07:06.698441] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:21:13.201 [2024-07-15 18:07:06.698525] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:21:13.201 [2024-07-15 18:07:06.698632] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:21:13.201 [2024-07-15 18:07:06.698634] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 4 00:21:13.764 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:21:13.764 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@862 -- # return 0 00:21:13.764 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:21:13.764 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@728 -- # xtrace_disable 00:21:13.764 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:21:13.764 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:21:13.764 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@20 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:21:13.764 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:13.764 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:21:13.764 [2024-07-15 18:07:07.408190] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:21:13.764 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:13.764 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@22 -- # num_subsystems=({1..10}) 00:21:13.764 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@24 -- # timing_enter create_subsystems 00:21:13.764 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@722 -- # xtrace_disable 00:21:13.764 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:21:13.764 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@26 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:21:13.764 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:13.764 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # cat 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # cat 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # cat 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # cat 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # cat 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # cat 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # cat 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # cat 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # cat 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@27 -- # for i in "${num_subsystems[@]}" 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@28 -- # cat 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@35 -- # rpc_cmd 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:13.765 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:21:13.765 Malloc1 00:21:14.022 [2024-07-15 18:07:07.500067] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:21:14.022 Malloc2 00:21:14.022 Malloc3 00:21:14.022 Malloc4 00:21:14.022 Malloc5 00:21:14.022 Malloc6 00:21:14.022 Malloc7 00:21:14.279 Malloc8 00:21:14.279 Malloc9 00:21:14.279 Malloc10 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@36 -- # timing_exit create_subsystems 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@728 -- # xtrace_disable 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@125 -- # perfpid=652222 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@126 -- # waitforlisten 652222 /var/tmp/bdevperf.sock 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@829 -- # '[' -z 652222 ']' 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@834 -- # local max_retries=100 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@124 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -r /var/tmp/bdevperf.sock --json /dev/fd/63 -q 64 -o 65536 -w verify -t 10 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@124 -- # gen_nvmf_target_json 1 2 3 4 5 6 7 8 9 10 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:21:14.279 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@838 -- # xtrace_disable 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@532 -- # config=() 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@532 -- # local subsystem config 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:21:14.279 { 00:21:14.279 "params": { 00:21:14.279 "name": "Nvme$subsystem", 00:21:14.279 "trtype": "$TEST_TRANSPORT", 00:21:14.279 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:14.279 "adrfam": "ipv4", 00:21:14.279 "trsvcid": "$NVMF_PORT", 00:21:14.279 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:14.279 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:14.279 "hdgst": ${hdgst:-false}, 00:21:14.279 "ddgst": ${ddgst:-false} 00:21:14.279 }, 00:21:14.279 "method": "bdev_nvme_attach_controller" 00:21:14.279 } 00:21:14.279 EOF 00:21:14.279 )") 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@554 -- # cat 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:21:14.279 { 00:21:14.279 "params": { 00:21:14.279 "name": "Nvme$subsystem", 00:21:14.279 "trtype": "$TEST_TRANSPORT", 00:21:14.279 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:14.279 "adrfam": "ipv4", 00:21:14.279 "trsvcid": "$NVMF_PORT", 00:21:14.279 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:14.279 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:14.279 "hdgst": ${hdgst:-false}, 00:21:14.279 "ddgst": ${ddgst:-false} 00:21:14.279 }, 00:21:14.279 "method": "bdev_nvme_attach_controller" 00:21:14.279 } 00:21:14.279 EOF 00:21:14.279 )") 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@554 -- # cat 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:21:14.279 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:21:14.279 { 00:21:14.279 "params": { 00:21:14.279 "name": "Nvme$subsystem", 00:21:14.280 "trtype": "$TEST_TRANSPORT", 00:21:14.280 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:14.280 "adrfam": "ipv4", 00:21:14.280 "trsvcid": "$NVMF_PORT", 00:21:14.280 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:14.280 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:14.280 "hdgst": ${hdgst:-false}, 00:21:14.280 "ddgst": ${ddgst:-false} 00:21:14.280 }, 00:21:14.280 "method": "bdev_nvme_attach_controller" 00:21:14.280 } 00:21:14.280 EOF 00:21:14.280 )") 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@554 -- # cat 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:21:14.280 { 00:21:14.280 "params": { 00:21:14.280 "name": "Nvme$subsystem", 00:21:14.280 "trtype": "$TEST_TRANSPORT", 00:21:14.280 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:14.280 "adrfam": "ipv4", 00:21:14.280 "trsvcid": "$NVMF_PORT", 00:21:14.280 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:14.280 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:14.280 "hdgst": ${hdgst:-false}, 00:21:14.280 "ddgst": ${ddgst:-false} 00:21:14.280 }, 00:21:14.280 "method": "bdev_nvme_attach_controller" 00:21:14.280 } 00:21:14.280 EOF 00:21:14.280 )") 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@554 -- # cat 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:21:14.280 { 00:21:14.280 "params": { 00:21:14.280 "name": "Nvme$subsystem", 00:21:14.280 "trtype": "$TEST_TRANSPORT", 00:21:14.280 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:14.280 "adrfam": "ipv4", 00:21:14.280 "trsvcid": "$NVMF_PORT", 00:21:14.280 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:14.280 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:14.280 "hdgst": ${hdgst:-false}, 00:21:14.280 "ddgst": ${ddgst:-false} 00:21:14.280 }, 00:21:14.280 "method": "bdev_nvme_attach_controller" 00:21:14.280 } 00:21:14.280 EOF 00:21:14.280 )") 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@554 -- # cat 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:21:14.280 { 00:21:14.280 "params": { 00:21:14.280 "name": "Nvme$subsystem", 00:21:14.280 "trtype": "$TEST_TRANSPORT", 00:21:14.280 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:14.280 "adrfam": "ipv4", 00:21:14.280 "trsvcid": "$NVMF_PORT", 00:21:14.280 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:14.280 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:14.280 "hdgst": ${hdgst:-false}, 00:21:14.280 "ddgst": ${ddgst:-false} 00:21:14.280 }, 00:21:14.280 "method": "bdev_nvme_attach_controller" 00:21:14.280 } 00:21:14.280 EOF 00:21:14.280 )") 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@554 -- # cat 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:21:14.280 [2024-07-15 18:07:07.966308] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:21:14.280 [2024-07-15 18:07:07.966359] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid652222 ] 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:21:14.280 { 00:21:14.280 "params": { 00:21:14.280 "name": "Nvme$subsystem", 00:21:14.280 "trtype": "$TEST_TRANSPORT", 00:21:14.280 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:14.280 "adrfam": "ipv4", 00:21:14.280 "trsvcid": "$NVMF_PORT", 00:21:14.280 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:14.280 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:14.280 "hdgst": ${hdgst:-false}, 00:21:14.280 "ddgst": ${ddgst:-false} 00:21:14.280 }, 00:21:14.280 "method": "bdev_nvme_attach_controller" 00:21:14.280 } 00:21:14.280 EOF 00:21:14.280 )") 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@554 -- # cat 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:21:14.280 { 00:21:14.280 "params": { 00:21:14.280 "name": "Nvme$subsystem", 00:21:14.280 "trtype": "$TEST_TRANSPORT", 00:21:14.280 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:14.280 "adrfam": "ipv4", 00:21:14.280 "trsvcid": "$NVMF_PORT", 00:21:14.280 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:14.280 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:14.280 "hdgst": ${hdgst:-false}, 00:21:14.280 "ddgst": ${ddgst:-false} 00:21:14.280 }, 00:21:14.280 "method": "bdev_nvme_attach_controller" 00:21:14.280 } 00:21:14.280 EOF 00:21:14.280 )") 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@554 -- # cat 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:21:14.280 { 00:21:14.280 "params": { 00:21:14.280 "name": "Nvme$subsystem", 00:21:14.280 "trtype": "$TEST_TRANSPORT", 00:21:14.280 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:14.280 "adrfam": "ipv4", 00:21:14.280 "trsvcid": "$NVMF_PORT", 00:21:14.280 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:14.280 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:14.280 "hdgst": ${hdgst:-false}, 00:21:14.280 "ddgst": ${ddgst:-false} 00:21:14.280 }, 00:21:14.280 "method": "bdev_nvme_attach_controller" 00:21:14.280 } 00:21:14.280 EOF 00:21:14.280 )") 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@554 -- # cat 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:21:14.280 { 00:21:14.280 "params": { 00:21:14.280 "name": "Nvme$subsystem", 00:21:14.280 "trtype": "$TEST_TRANSPORT", 00:21:14.280 "traddr": "$NVMF_FIRST_TARGET_IP", 00:21:14.280 "adrfam": "ipv4", 00:21:14.280 "trsvcid": "$NVMF_PORT", 00:21:14.280 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:21:14.280 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:21:14.280 "hdgst": ${hdgst:-false}, 00:21:14.280 "ddgst": ${ddgst:-false} 00:21:14.280 }, 00:21:14.280 "method": "bdev_nvme_attach_controller" 00:21:14.280 } 00:21:14.280 EOF 00:21:14.280 )") 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@554 -- # cat 00:21:14.280 EAL: No free 2048 kB hugepages reported on node 1 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@556 -- # jq . 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@557 -- # IFS=, 00:21:14.280 18:07:07 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:21:14.280 "params": { 00:21:14.280 "name": "Nvme1", 00:21:14.280 "trtype": "tcp", 00:21:14.280 "traddr": "10.0.0.2", 00:21:14.280 "adrfam": "ipv4", 00:21:14.280 "trsvcid": "4420", 00:21:14.280 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:21:14.280 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:21:14.280 "hdgst": false, 00:21:14.280 "ddgst": false 00:21:14.280 }, 00:21:14.280 "method": "bdev_nvme_attach_controller" 00:21:14.280 },{ 00:21:14.280 "params": { 00:21:14.280 "name": "Nvme2", 00:21:14.280 "trtype": "tcp", 00:21:14.280 "traddr": "10.0.0.2", 00:21:14.280 "adrfam": "ipv4", 00:21:14.280 "trsvcid": "4420", 00:21:14.280 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:21:14.280 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:21:14.280 "hdgst": false, 00:21:14.280 "ddgst": false 00:21:14.280 }, 00:21:14.280 "method": "bdev_nvme_attach_controller" 00:21:14.280 },{ 00:21:14.280 "params": { 00:21:14.280 "name": "Nvme3", 00:21:14.280 "trtype": "tcp", 00:21:14.280 "traddr": "10.0.0.2", 00:21:14.280 "adrfam": "ipv4", 00:21:14.280 "trsvcid": "4420", 00:21:14.280 "subnqn": "nqn.2016-06.io.spdk:cnode3", 00:21:14.280 "hostnqn": "nqn.2016-06.io.spdk:host3", 00:21:14.280 "hdgst": false, 00:21:14.280 "ddgst": false 00:21:14.280 }, 00:21:14.280 "method": "bdev_nvme_attach_controller" 00:21:14.280 },{ 00:21:14.281 "params": { 00:21:14.281 "name": "Nvme4", 00:21:14.281 "trtype": "tcp", 00:21:14.281 "traddr": "10.0.0.2", 00:21:14.281 "adrfam": "ipv4", 00:21:14.281 "trsvcid": "4420", 00:21:14.281 "subnqn": "nqn.2016-06.io.spdk:cnode4", 00:21:14.281 "hostnqn": "nqn.2016-06.io.spdk:host4", 00:21:14.281 "hdgst": false, 00:21:14.281 "ddgst": false 00:21:14.281 }, 00:21:14.281 "method": "bdev_nvme_attach_controller" 00:21:14.281 },{ 00:21:14.281 "params": { 00:21:14.281 "name": "Nvme5", 00:21:14.281 "trtype": "tcp", 00:21:14.281 "traddr": "10.0.0.2", 00:21:14.281 "adrfam": "ipv4", 00:21:14.281 "trsvcid": "4420", 00:21:14.281 "subnqn": "nqn.2016-06.io.spdk:cnode5", 00:21:14.281 "hostnqn": "nqn.2016-06.io.spdk:host5", 00:21:14.281 "hdgst": false, 00:21:14.281 "ddgst": false 00:21:14.281 }, 00:21:14.281 "method": "bdev_nvme_attach_controller" 00:21:14.281 },{ 00:21:14.281 "params": { 00:21:14.281 "name": "Nvme6", 00:21:14.281 "trtype": "tcp", 00:21:14.281 "traddr": "10.0.0.2", 00:21:14.281 "adrfam": "ipv4", 00:21:14.281 "trsvcid": "4420", 00:21:14.281 "subnqn": "nqn.2016-06.io.spdk:cnode6", 00:21:14.281 "hostnqn": "nqn.2016-06.io.spdk:host6", 00:21:14.281 "hdgst": false, 00:21:14.281 "ddgst": false 00:21:14.281 }, 00:21:14.281 "method": "bdev_nvme_attach_controller" 00:21:14.281 },{ 00:21:14.281 "params": { 00:21:14.281 "name": "Nvme7", 00:21:14.281 "trtype": "tcp", 00:21:14.281 "traddr": "10.0.0.2", 00:21:14.281 "adrfam": "ipv4", 00:21:14.281 "trsvcid": "4420", 00:21:14.281 "subnqn": "nqn.2016-06.io.spdk:cnode7", 00:21:14.281 "hostnqn": "nqn.2016-06.io.spdk:host7", 00:21:14.281 "hdgst": false, 00:21:14.281 "ddgst": false 00:21:14.281 }, 00:21:14.281 "method": "bdev_nvme_attach_controller" 00:21:14.281 },{ 00:21:14.281 "params": { 00:21:14.281 "name": "Nvme8", 00:21:14.281 "trtype": "tcp", 00:21:14.281 "traddr": "10.0.0.2", 00:21:14.281 "adrfam": "ipv4", 00:21:14.281 "trsvcid": "4420", 00:21:14.281 "subnqn": "nqn.2016-06.io.spdk:cnode8", 00:21:14.281 "hostnqn": "nqn.2016-06.io.spdk:host8", 00:21:14.281 "hdgst": false, 00:21:14.281 "ddgst": false 00:21:14.281 }, 00:21:14.281 "method": "bdev_nvme_attach_controller" 00:21:14.281 },{ 00:21:14.281 "params": { 00:21:14.281 "name": "Nvme9", 00:21:14.281 "trtype": "tcp", 00:21:14.281 "traddr": "10.0.0.2", 00:21:14.281 "adrfam": "ipv4", 00:21:14.281 "trsvcid": "4420", 00:21:14.281 "subnqn": "nqn.2016-06.io.spdk:cnode9", 00:21:14.281 "hostnqn": "nqn.2016-06.io.spdk:host9", 00:21:14.281 "hdgst": false, 00:21:14.281 "ddgst": false 00:21:14.281 }, 00:21:14.281 "method": "bdev_nvme_attach_controller" 00:21:14.281 },{ 00:21:14.281 "params": { 00:21:14.281 "name": "Nvme10", 00:21:14.281 "trtype": "tcp", 00:21:14.281 "traddr": "10.0.0.2", 00:21:14.281 "adrfam": "ipv4", 00:21:14.281 "trsvcid": "4420", 00:21:14.281 "subnqn": "nqn.2016-06.io.spdk:cnode10", 00:21:14.281 "hostnqn": "nqn.2016-06.io.spdk:host10", 00:21:14.281 "hdgst": false, 00:21:14.281 "ddgst": false 00:21:14.281 }, 00:21:14.281 "method": "bdev_nvme_attach_controller" 00:21:14.281 }' 00:21:14.537 [2024-07-15 18:07:08.023102] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:14.537 [2024-07-15 18:07:08.096335] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:21:16.430 Running I/O for 10 seconds... 00:21:16.430 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:21:16.430 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@862 -- # return 0 00:21:16.430 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@127 -- # rpc_cmd -s /var/tmp/bdevperf.sock framework_wait_init 00:21:16.430 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:16.430 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:21:16.430 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:16.430 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@130 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill -9 $perfpid || true; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:21:16.431 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@132 -- # waitforio /var/tmp/bdevperf.sock Nvme1n1 00:21:16.431 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@50 -- # '[' -z /var/tmp/bdevperf.sock ']' 00:21:16.431 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@54 -- # '[' -z Nvme1n1 ']' 00:21:16.431 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@57 -- # local ret=1 00:21:16.431 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@58 -- # local i 00:21:16.431 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@59 -- # (( i = 10 )) 00:21:16.431 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@59 -- # (( i != 0 )) 00:21:16.431 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:21:16.431 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:16.431 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # jq -r '.bdevs[0].num_read_ops' 00:21:16.431 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:21:16.431 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:16.431 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # read_io_count=3 00:21:16.431 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@63 -- # '[' 3 -ge 100 ']' 00:21:16.431 18:07:09 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@67 -- # sleep 0.25 00:21:16.686 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@59 -- # (( i-- )) 00:21:16.686 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@59 -- # (( i != 0 )) 00:21:16.686 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:21:16.686 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # jq -r '.bdevs[0].num_read_ops' 00:21:16.686 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:16.686 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:21:16.686 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:16.686 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # read_io_count=67 00:21:16.686 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@63 -- # '[' 67 -ge 100 ']' 00:21:16.686 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@67 -- # sleep 0.25 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@59 -- # (( i-- )) 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@59 -- # (( i != 0 )) 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_get_iostat -b Nvme1n1 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # jq -r '.bdevs[0].num_read_ops' 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@60 -- # read_io_count=131 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@63 -- # '[' 131 -ge 100 ']' 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@64 -- # ret=0 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@65 -- # break 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@69 -- # return 0 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@135 -- # killprocess 651945 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@948 -- # '[' -z 651945 ']' 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@952 -- # kill -0 651945 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@953 -- # uname 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 651945 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@966 -- # echo 'killing process with pid 651945' 00:21:16.958 killing process with pid 651945 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@967 -- # kill 651945 00:21:16.958 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@972 -- # wait 651945 00:21:16.958 [2024-07-15 18:07:10.567427] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2376950 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.568048] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377f60 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.568077] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377f60 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.568085] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377f60 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.568099] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377f60 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.568106] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377f60 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.568113] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377f60 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.568119] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377f60 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.568126] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377f60 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.568132] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377f60 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.568139] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377f60 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.568145] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377f60 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.569100] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2376df0 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570333] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570357] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570365] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570372] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570379] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570386] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570393] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570400] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570407] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570414] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570421] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570427] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570433] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570440] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570447] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570454] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570461] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570467] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570479] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570485] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570492] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570499] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570506] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570512] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570519] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570526] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570532] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570539] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570546] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570552] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570559] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570566] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570573] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.958 [2024-07-15 18:07:10.570578] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570585] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570592] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570598] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570604] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570612] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570618] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570625] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570631] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570638] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570644] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570651] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570659] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570665] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570671] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570678] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570684] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570691] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570697] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570703] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570709] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570715] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570721] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570728] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570734] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570740] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570746] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570753] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570759] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.570766] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377290 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.571914] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.571938] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.571946] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.571953] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.571961] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.571968] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.571975] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.571982] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.571988] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.571998] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572005] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572012] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572018] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572024] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572030] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572036] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572043] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572049] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572057] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572063] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572069] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572076] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572082] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572089] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572095] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572101] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572108] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572115] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572121] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572127] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572133] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572140] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572146] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572152] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572160] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572166] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572174] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572181] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572187] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572193] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572199] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572206] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572212] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572218] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572228] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572235] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572241] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572249] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572255] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572262] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572269] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572275] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572282] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572288] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572294] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572300] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572306] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572313] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572320] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572326] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.959 [2024-07-15 18:07:10.572333] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.572339] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.572345] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x2377750 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573316] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573335] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573343] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573350] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573357] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573364] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573371] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573377] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573384] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573390] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573396] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573403] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573410] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573416] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573423] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573430] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573435] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573442] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573448] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573454] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573461] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573467] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573474] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573480] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573486] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573493] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573499] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573505] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573516] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573522] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573529] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573536] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573542] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573549] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573556] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573562] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573568] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573574] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573580] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573587] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573594] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573601] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573607] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573613] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573619] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573625] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573631] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573637] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573644] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573651] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573657] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573663] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573669] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573676] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573682] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573691] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573697] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573704] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573711] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573717] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573723] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573729] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.573736] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fa940 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574646] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574659] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574666] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574673] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574680] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574686] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574693] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574699] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574706] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574712] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574719] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574726] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574732] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574739] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574745] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574751] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574757] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574764] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574770] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574780] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574786] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574792] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574799] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574805] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574811] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574818] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.960 [2024-07-15 18:07:10.574825] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574831] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574837] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574843] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574849] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574855] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574862] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574869] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574875] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574881] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574888] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574894] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574900] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574906] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574913] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574919] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574925] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574931] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574937] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574944] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574951] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574957] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574963] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574971] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574978] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574984] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574991] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.574997] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.575004] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.575010] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.575016] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.575022] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.575028] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.575034] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.575040] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.575047] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.575053] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fade0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576029] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576044] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576050] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576057] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576063] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576070] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576076] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576082] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576088] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576094] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576103] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576109] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576115] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576122] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576128] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576134] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576140] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576147] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576153] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576159] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576164] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576171] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576176] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576182] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576189] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576195] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576201] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576207] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576213] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576221] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576234] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576241] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576247] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576254] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576260] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576266] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576272] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576281] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576287] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576293] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576299] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576305] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576311] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576317] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576323] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576329] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576335] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576341] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576348] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576354] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576360] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576365] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576371] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576378] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.961 [2024-07-15 18:07:10.576383] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.576389] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.576396] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.576402] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.576407] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.576413] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.576419] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.576425] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.576431] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb2a0 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577181] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577196] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577202] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577208] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577214] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577220] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577230] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577237] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577243] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577249] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577255] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577261] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577267] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577273] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577280] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577286] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577292] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577298] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577304] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577310] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577316] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577323] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577329] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577335] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577341] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577347] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577353] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577359] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577366] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577372] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577378] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577385] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577391] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577397] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577404] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577410] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577416] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577423] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577429] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577435] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577441] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577447] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577454] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577459] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577466] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577472] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577478] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577485] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577490] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577496] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577502] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577508] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577514] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577520] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577526] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577533] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577540] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577547] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577552] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577559] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577565] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577571] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.577578] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x23fb740 is same with the state(5) to be set 00:21:16.962 [2024-07-15 18:07:10.580369] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.962 [2024-07-15 18:07:10.580400] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.962 [2024-07-15 18:07:10.580410] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.962 [2024-07-15 18:07:10.580418] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.962 [2024-07-15 18:07:10.580426] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580433] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580441] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580448] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580456] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1564b30 is same with the state(5) to be set 00:21:16.963 [2024-07-15 18:07:10.580484] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580493] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580500] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580515] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580522] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580530] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580537] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580544] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x16d58b0 is same with the state(5) to be set 00:21:16.963 [2024-07-15 18:07:10.580569] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580589] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580595] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580603] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580616] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580623] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580630] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x16ec0d0 is same with the state(5) to be set 00:21:16.963 [2024-07-15 18:07:10.580656] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580664] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580671] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580678] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580685] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580691] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580698] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580705] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580712] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x155d1d0 is same with the state(5) to be set 00:21:16.963 [2024-07-15 18:07:10.580734] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580742] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580750] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580757] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580764] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580770] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580778] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580785] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580797] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x106f340 is same with the state(5) to be set 00:21:16.963 [2024-07-15 18:07:10.580820] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580837] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580845] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580852] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580860] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580867] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580874] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580881] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x16f5050 is same with the state(5) to be set 00:21:16.963 [2024-07-15 18:07:10.580903] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580912] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580919] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580926] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580934] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580942] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580950] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580956] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580963] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x16ec8d0 is same with the state(5) to be set 00:21:16.963 [2024-07-15 18:07:10.580983] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.580991] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.580999] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.581006] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.581015] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.581022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.581029] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.581038] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.581045] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1567bf0 is same with the state(5) to be set 00:21:16.963 [2024-07-15 18:07:10.581068] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.581076] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.581085] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.581091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.581099] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.581106] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.581113] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.581120] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.581127] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1543190 is same with the state(5) to be set 00:21:16.963 [2024-07-15 18:07:10.581150] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.581158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.581166] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.581173] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.581181] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.581188] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.581196] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:16.963 [2024-07-15 18:07:10.581203] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.581209] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1520c70 is same with the state(5) to be set 00:21:16.963 [2024-07-15 18:07:10.581285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.963 [2024-07-15 18:07:10.581296] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.963 [2024-07-15 18:07:10.581309] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581325] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581336] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581345] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581352] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581360] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581367] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581376] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581384] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581393] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581400] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581408] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581415] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581423] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581445] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581454] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581460] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581469] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581484] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581491] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581500] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581507] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581515] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581523] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581533] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581540] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581548] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581557] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581566] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581582] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581597] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581612] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581619] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581628] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581634] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581643] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581650] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581665] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581674] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581681] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581689] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581699] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581708] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581715] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581723] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581740] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581747] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581756] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581772] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581787] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581803] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581823] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581839] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581846] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581854] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581862] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581870] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581877] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581885] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581892] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581908] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581924] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581934] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581941] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581950] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581956] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581972] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.964 [2024-07-15 18:07:10.581981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.964 [2024-07-15 18:07:10.581987] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.581996] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582002] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582011] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582018] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582027] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582033] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582042] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582048] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582056] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582065] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582073] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582081] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582096] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582105] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582112] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582121] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582138] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582144] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582153] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582162] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582186] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582201] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582208] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582217] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582228] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582237] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582246] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582255] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582262] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582271] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582278] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582293] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582301] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582308] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582370] bdev_nvme.c:1612:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x1661ea0 was disconnected and freed. reset controller. 00:21:16.965 [2024-07-15 18:07:10.582736] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582758] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582778] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582786] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582793] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582801] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582809] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582817] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582832] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582839] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582855] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582863] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582870] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582879] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582886] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582895] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582903] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582911] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582918] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582926] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582941] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582948] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582958] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582966] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582975] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582982] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.582990] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.582997] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.583007] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.583015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.583023] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.583030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.583039] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.583046] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.583055] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.583062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.583070] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.583078] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.583086] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.965 [2024-07-15 18:07:10.583093] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.965 [2024-07-15 18:07:10.583102] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.583108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.583117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.583124] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.583132] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.583139] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.583147] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.583156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.583164] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.583171] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.583179] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.583186] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.583194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.583202] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.583210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.583217] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595086] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595106] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595115] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595122] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595132] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595139] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595163] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595169] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595178] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595184] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595193] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595199] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595208] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595215] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595230] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595238] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595246] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595253] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595268] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595277] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595283] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595291] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595298] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595306] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595312] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595320] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595327] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595335] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595342] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595350] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595357] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595365] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595380] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595386] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595410] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595418] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595426] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595433] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595442] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595449] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595457] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595464] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595472] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595493] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595501] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595516] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595522] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595537] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.966 [2024-07-15 18:07:10.595545] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.966 [2024-07-15 18:07:10.595551] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.595560] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.595566] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.595575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.595581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.595589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.595596] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.595605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.595613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.595650] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:21:16.967 [2024-07-15 18:07:10.595705] bdev_nvme.c:1612:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x15e8490 was disconnected and freed. reset controller. 00:21:16.967 [2024-07-15 18:07:10.596136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596179] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596188] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596195] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596203] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596210] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596218] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596232] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596241] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596248] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596257] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596263] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596272] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596278] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596287] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596293] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596301] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596308] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596316] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596322] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596334] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596340] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596348] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596355] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596364] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596379] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596386] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596401] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596410] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596416] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596424] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596445] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596460] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596468] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596474] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596482] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596489] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596497] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596503] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596512] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596520] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596529] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596535] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596543] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596558] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596564] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596573] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596594] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596602] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596617] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596624] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596633] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596647] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596653] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596668] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596676] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596683] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596691] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596698] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596718] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596724] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.967 [2024-07-15 18:07:10.596733] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.967 [2024-07-15 18:07:10.596739] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.596747] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.596754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.596762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.596768] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.596776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.596783] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.596791] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.596798] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.596807] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.596814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.596822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.596828] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.596836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.596843] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.596851] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.596857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.596865] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.596872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.596881] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.596887] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.596895] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.596903] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.596912] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.596918] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.596926] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.596932] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.596941] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.596947] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.596955] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.596961] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.596969] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.596976] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.596984] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.596990] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.596999] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.597005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.597013] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.597020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.597028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.597035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.597043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.597049] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.597057] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.597064] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.597072] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.597078] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.597088] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.597108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.597120] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.597130] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.597141] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.968 [2024-07-15 18:07:10.597150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.968 [2024-07-15 18:07:10.600495] bdev_nvme.c:1612:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x164aef0 was disconnected and freed. reset controller. 00:21:16.968 [2024-07-15 18:07:10.600570] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1564b30 (9): Bad file descriptor 00:21:16.968 [2024-07-15 18:07:10.600595] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x16d58b0 (9): Bad file descriptor 00:21:16.968 [2024-07-15 18:07:10.600610] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x16ec0d0 (9): Bad file descriptor 00:21:16.968 [2024-07-15 18:07:10.600625] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x155d1d0 (9): Bad file descriptor 00:21:16.968 [2024-07-15 18:07:10.600644] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x106f340 (9): Bad file descriptor 00:21:16.968 [2024-07-15 18:07:10.600663] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x16f5050 (9): Bad file descriptor 00:21:16.968 [2024-07-15 18:07:10.600681] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x16ec8d0 (9): Bad file descriptor 00:21:16.968 [2024-07-15 18:07:10.600696] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1567bf0 (9): Bad file descriptor 00:21:16.968 [2024-07-15 18:07:10.600710] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1543190 (9): Bad file descriptor 00:21:16.968 [2024-07-15 18:07:10.600726] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1520c70 (9): Bad file descriptor 00:21:16.968 [2024-07-15 18:07:10.604829] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:21:16.968 [2024-07-15 18:07:10.604868] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode2] resetting controller 00:21:16.968 [2024-07-15 18:07:10.605342] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode10] resetting controller 00:21:16.968 [2024-07-15 18:07:10.605549] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:16.968 [2024-07-15 18:07:10.605569] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1520c70 with addr=10.0.0.2, port=4420 00:21:16.968 [2024-07-15 18:07:10.605580] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1520c70 is same with the state(5) to be set 00:21:16.968 [2024-07-15 18:07:10.605722] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:16.968 [2024-07-15 18:07:10.605735] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x16ec8d0 with addr=10.0.0.2, port=4420 00:21:16.968 [2024-07-15 18:07:10.605745] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x16ec8d0 is same with the state(5) to be set 00:21:16.968 [2024-07-15 18:07:10.608827] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:16.968 [2024-07-15 18:07:10.608857] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x16d58b0 with addr=10.0.0.2, port=4420 00:21:16.968 [2024-07-15 18:07:10.608872] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x16d58b0 is same with the state(5) to be set 00:21:16.968 [2024-07-15 18:07:10.608887] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1520c70 (9): Bad file descriptor 00:21:16.968 [2024-07-15 18:07:10.608900] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x16ec8d0 (9): Bad file descriptor 00:21:16.968 [2024-07-15 18:07:10.608974] nvme_tcp.c:1241:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:21:16.968 [2024-07-15 18:07:10.609029] nvme_tcp.c:1241:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:21:16.968 [2024-07-15 18:07:10.609079] nvme_tcp.c:1241:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:21:16.968 [2024-07-15 18:07:10.609132] nvme_tcp.c:1241:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:21:16.968 [2024-07-15 18:07:10.609182] nvme_tcp.c:1241:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:21:16.968 [2024-07-15 18:07:10.609243] nvme_tcp.c:1241:nvme_tcp_pdu_ch_handle: *ERROR*: Unexpected PDU type 0x00 00:21:16.968 [2024-07-15 18:07:10.609347] bdev_nvme.c:1612:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x1649a60 was disconnected and freed. reset controller. 00:21:16.968 [2024-07-15 18:07:10.609385] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x16d58b0 (9): Bad file descriptor 00:21:16.968 [2024-07-15 18:07:10.609398] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:21:16.968 [2024-07-15 18:07:10.609407] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:21:16.968 [2024-07-15 18:07:10.609416] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:21:16.968 [2024-07-15 18:07:10.609432] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode2] Ctrlr is in error state 00:21:16.968 [2024-07-15 18:07:10.609440] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode2] controller reinitialization failed 00:21:16.969 [2024-07-15 18:07:10.609449] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode2] in failed state. 00:21:16.969 [2024-07-15 18:07:10.609875] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:16.969 [2024-07-15 18:07:10.609889] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:16.969 [2024-07-15 18:07:10.609898] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode9] resetting controller 00:21:16.969 [2024-07-15 18:07:10.609919] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode10] Ctrlr is in error state 00:21:16.969 [2024-07-15 18:07:10.609928] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode10] controller reinitialization failed 00:21:16.969 [2024-07-15 18:07:10.609936] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode10] in failed state. 00:21:16.969 [2024-07-15 18:07:10.609993] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:16.969 [2024-07-15 18:07:10.610129] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:16.969 [2024-07-15 18:07:10.610145] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x16ec0d0 with addr=10.0.0.2, port=4420 00:21:16.969 [2024-07-15 18:07:10.610154] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x16ec0d0 is same with the state(5) to be set 00:21:16.969 [2024-07-15 18:07:10.610203] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x16ec0d0 (9): Bad file descriptor 00:21:16.969 [2024-07-15 18:07:10.610250] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode9] Ctrlr is in error state 00:21:16.969 [2024-07-15 18:07:10.610260] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode9] controller reinitialization failed 00:21:16.969 [2024-07-15 18:07:10.610268] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode9] in failed state. 00:21:16.969 [2024-07-15 18:07:10.610315] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:16.969 [2024-07-15 18:07:10.610685] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.610699] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.610714] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.610723] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.610735] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.610744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.610755] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.610764] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.610775] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.610784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.610795] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.610804] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.610815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.610824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.610836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.610844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.610855] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.610864] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.610875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.610884] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.610896] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.610905] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.610916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.610925] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.610939] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.610948] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.610959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.610968] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.610979] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.610988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.610999] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611008] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611018] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611027] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611047] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611058] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611067] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611097] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611106] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611126] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611145] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611157] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611166] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611177] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611198] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611208] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611219] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611246] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611257] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611277] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611328] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611337] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611347] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611357] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611368] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611377] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611388] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611396] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.969 [2024-07-15 18:07:10.611416] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.969 [2024-07-15 18:07:10.611427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611436] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611458] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611469] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611478] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611508] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611518] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611529] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611549] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611558] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611569] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611599] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611610] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611630] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611650] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611659] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611669] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611678] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611689] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611700] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611720] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611731] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611741] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611752] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611780] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611791] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611800] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611811] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611820] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611830] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611840] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611851] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611860] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611871] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611891] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611910] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611919] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611930] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611939] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611950] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611961] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611972] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.611980] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.611990] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x15e9920 is same with the state(5) to be set 00:21:16.970 [2024-07-15 18:07:10.613359] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.613373] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.613386] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.613396] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.613407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.613416] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.613427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.613436] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.613448] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.613456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.613467] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.613476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.613487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.613496] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.613507] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.613516] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.970 [2024-07-15 18:07:10.613526] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.970 [2024-07-15 18:07:10.613535] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613555] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613566] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613578] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613590] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613599] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613611] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613631] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613651] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613660] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613671] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613680] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613691] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613699] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613719] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613730] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613739] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613750] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613758] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613778] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613797] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613808] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613817] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613830] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613839] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613859] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613869] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613879] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613890] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613898] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613909] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613918] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613929] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613939] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613950] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613959] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613970] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613979] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.613990] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.613999] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.614010] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.614018] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.614029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.614038] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.614049] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.614058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.614069] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.614079] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.614090] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.614099] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.614110] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.614119] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.614130] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.614139] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.614150] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.614159] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.614170] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.614179] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.614190] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.614199] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.614210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.614219] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.614238] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.614248] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.614259] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.614268] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.614279] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.614288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.614299] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.614308] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.614319] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.614328] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.614341] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.614350] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.614361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.614370] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.614382] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.971 [2024-07-15 18:07:10.614391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.971 [2024-07-15 18:07:10.614402] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.614411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.614422] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.614431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.614442] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.614451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.614462] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.614471] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.614482] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.614491] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.614502] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.614511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.614522] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.614531] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.614542] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.614550] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.614561] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.614571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.614582] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.614593] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.614604] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.614613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.614624] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.614633] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.614644] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.614653] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.614662] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x151ab70 is same with the state(5) to be set 00:21:16.972 [2024-07-15 18:07:10.616013] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616027] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616039] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616048] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616059] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616068] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616080] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616088] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616099] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616118] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616127] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616139] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616158] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616167] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616178] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616201] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616210] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616220] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616245] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616265] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616274] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616284] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616293] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616304] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616313] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616346] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616357] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616367] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616378] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616387] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616398] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616418] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616447] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616462] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616471] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616482] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616491] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616502] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616522] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616531] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616542] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616551] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616583] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616592] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616602] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616612] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.972 [2024-07-15 18:07:10.616622] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.972 [2024-07-15 18:07:10.616631] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.616642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.616651] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.616662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.616671] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.616682] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.616691] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.616702] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.616714] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.616725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.616734] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.616745] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.616753] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.616764] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.616773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.616784] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.616793] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.616804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.616812] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.616823] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.616832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.616844] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.616853] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.616863] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.616872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.616883] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.616892] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.616903] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.616911] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.616922] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.616931] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.616942] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.616951] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.616963] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.616972] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.616984] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.616992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.617003] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.617012] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.617023] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.617032] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.617043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.617052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.617063] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.617071] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.617082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.617091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.617102] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.617111] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.617122] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.617131] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.617152] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.617158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.617169] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.617175] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.617183] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.617190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.617198] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.617205] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.617214] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.617220] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.617231] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.617238] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.617246] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.617252] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.617261] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.617267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.617275] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.617282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.617289] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x151c040 is same with the state(5) to be set 00:21:16.973 [2024-07-15 18:07:10.618311] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.618322] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.618332] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.618338] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.618347] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.618353] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.618362] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.618368] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.618377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.618383] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.618392] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.618399] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.618407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.618416] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.973 [2024-07-15 18:07:10.618424] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.973 [2024-07-15 18:07:10.618431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618455] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618461] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618469] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618475] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618484] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618490] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618498] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618512] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618519] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618527] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618533] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618541] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618548] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618570] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618585] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618591] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618600] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618607] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618616] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618622] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618630] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618637] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618645] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618651] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618659] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618666] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618674] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618680] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618688] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618703] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618710] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618724] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618732] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618738] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618746] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618753] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618775] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618782] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618791] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618798] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618806] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618812] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618821] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618827] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618835] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618849] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618855] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618864] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618871] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618879] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618885] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618893] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618908] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618922] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618929] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618937] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618943] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618951] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618957] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.974 [2024-07-15 18:07:10.618973] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.974 [2024-07-15 18:07:10.618981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.618988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.618995] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.619002] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.619010] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.619016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.619024] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.619030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.619038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.619045] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.619053] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.619060] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.619068] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.619074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.619082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.619088] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.619096] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.619103] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.619111] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.619118] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.619126] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.619132] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.619140] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.619147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.619156] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.619163] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.619171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.619177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.619186] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.619192] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.619200] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.619206] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.619214] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.619220] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.619232] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.619238] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.619246] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.619253] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.619260] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1662910 is same with the state(5) to be set 00:21:16.975 [2024-07-15 18:07:10.620278] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620290] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620300] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620307] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620315] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620322] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620330] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620337] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620345] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620352] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620362] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620369] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620384] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620393] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620411] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620418] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620427] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620433] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620442] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620448] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620456] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620471] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620478] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620502] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620517] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620524] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620532] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620539] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620547] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620555] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620564] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620570] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620578] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620585] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620593] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620609] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620623] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.975 [2024-07-15 18:07:10.620630] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.975 [2024-07-15 18:07:10.620638] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620645] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620654] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620660] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620668] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620675] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620683] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620698] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620705] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620713] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620720] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620735] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620744] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620751] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620759] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620766] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620774] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620781] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620803] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620810] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620818] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620824] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620833] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620840] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620855] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620864] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620870] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620878] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620885] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620893] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620908] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620922] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620930] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620938] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620945] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620953] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620960] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620974] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620982] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.620989] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.620997] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.621003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.621011] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.621018] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.621026] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.621033] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.621041] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.621048] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.621056] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.621062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.621070] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.621077] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.621085] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.621091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.621099] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.621106] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.621115] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.621122] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.621130] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.621136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.621144] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.621151] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.621159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.621165] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.621173] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.621180] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.621188] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.621194] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.621202] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.621208] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.621217] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.621223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.621234] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.621241] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.621248] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1663de0 is same with the state(5) to be set 00:21:16.976 [2024-07-15 18:07:10.622278] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.622291] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.976 [2024-07-15 18:07:10.622301] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24704 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.976 [2024-07-15 18:07:10.622308] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622317] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24832 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622325] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622341] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24960 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622348] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:25088 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622363] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622372] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:25216 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622378] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622386] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:25344 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622392] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622401] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:25472 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622415] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:25600 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622430] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:25728 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622437] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622445] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:25856 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622452] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622460] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:25984 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622466] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622475] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:26112 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:26240 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622496] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:26368 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622510] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622518] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:26496 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622529] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622538] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:26624 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622552] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:26752 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622559] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:26880 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622583] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:27008 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622589] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622598] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:27136 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622613] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:27264 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622619] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622627] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:27392 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622634] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:27520 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622648] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622656] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:27648 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622663] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622672] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:27776 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622678] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622686] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:27904 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622701] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:28032 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622707] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:28160 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622723] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622731] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:28288 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622738] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622746] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:28416 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622760] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:28544 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622775] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:28672 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622781] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622790] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:28800 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622796] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:28928 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622810] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:29056 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622826] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622834] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:29184 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622849] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:29312 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622855] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622864] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:29440 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622870] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622878] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:29568 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622884] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:29696 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622908] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:29824 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:29952 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622929] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622938] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:30080 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622944] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.977 [2024-07-15 18:07:10.622952] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:30208 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.977 [2024-07-15 18:07:10.622959] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.978 [2024-07-15 18:07:10.622967] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:30336 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.978 [2024-07-15 18:07:10.622973] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.978 [2024-07-15 18:07:10.622982] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:30464 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.978 [2024-07-15 18:07:10.622989] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.978 [2024-07-15 18:07:10.622997] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:30592 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.978 [2024-07-15 18:07:10.623003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.978 [2024-07-15 18:07:10.623012] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:30720 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.978 [2024-07-15 18:07:10.623018] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.978 [2024-07-15 18:07:10.623026] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:30848 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.978 [2024-07-15 18:07:10.623032] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.978 [2024-07-15 18:07:10.623040] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:30976 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.978 [2024-07-15 18:07:10.623047] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.978 [2024-07-15 18:07:10.623055] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:31104 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.978 [2024-07-15 18:07:10.623062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.978 [2024-07-15 18:07:10.623070] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:31232 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.978 [2024-07-15 18:07:10.623077] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.978 [2024-07-15 18:07:10.623086] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:31360 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.978 [2024-07-15 18:07:10.623093] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.978 [2024-07-15 18:07:10.623101] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:31488 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.978 [2024-07-15 18:07:10.623107] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.978 [2024-07-15 18:07:10.623116] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:31616 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.978 [2024-07-15 18:07:10.623122] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.978 [2024-07-15 18:07:10.623130] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:31744 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.978 [2024-07-15 18:07:10.623136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.978 [2024-07-15 18:07:10.623145] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:31872 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.978 [2024-07-15 18:07:10.623151] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.978 [2024-07-15 18:07:10.623159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:32000 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.978 [2024-07-15 18:07:10.623166] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.978 [2024-07-15 18:07:10.623175] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:32128 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.978 [2024-07-15 18:07:10.623182] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.978 [2024-07-15 18:07:10.623190] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:32256 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.978 [2024-07-15 18:07:10.623196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.978 [2024-07-15 18:07:10.623205] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:32384 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.978 [2024-07-15 18:07:10.623211] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.978 [2024-07-15 18:07:10.623219] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:32512 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.978 [2024-07-15 18:07:10.623230] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.978 [2024-07-15 18:07:10.623239] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:32640 len:128 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:16.978 [2024-07-15 18:07:10.623245] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:16.978 [2024-07-15 18:07:10.623252] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x16652b0 is same with the state(5) to be set 00:21:16.978 [2024-07-15 18:07:10.624503] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode3] resetting controller 00:21:16.978 [2024-07-15 18:07:10.624524] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode4] resetting controller 00:21:16.978 [2024-07-15 18:07:10.624534] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode5] resetting controller 00:21:16.978 [2024-07-15 18:07:10.624545] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode6] resetting controller 00:21:16.978 [2024-07-15 18:07:10.624617] bdev_nvme.c:2899:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:21:16.978 [2024-07-15 18:07:10.624629] bdev_nvme.c:2899:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:21:16.978 [2024-07-15 18:07:10.624688] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode7] resetting controller 00:21:16.978 task offset: 31616 on job bdev=Nvme1n1 fails 00:21:16.978 00:21:16.978 Latency(us) 00:21:16.978 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:21:16.978 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:16.978 Job: Nvme1n1 ended in about 0.90 seconds with error 00:21:16.978 Verification LBA range: start 0x0 length 0x400 00:21:16.978 Nvme1n1 : 0.90 212.17 13.26 70.72 0.00 223969.28 18692.01 217921.45 00:21:16.978 Job: Nvme2n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:16.978 Job: Nvme2n1 ended in about 0.91 seconds with error 00:21:16.978 Verification LBA range: start 0x0 length 0x400 00:21:16.978 Nvme2n1 : 0.91 211.89 13.24 70.63 0.00 220261.29 18008.15 216097.84 00:21:16.978 Job: Nvme3n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:16.978 Job: Nvme3n1 ended in about 0.92 seconds with error 00:21:16.978 Verification LBA range: start 0x0 length 0x400 00:21:16.978 Nvme3n1 : 0.92 209.57 13.10 69.86 0.00 218775.82 13335.15 222480.47 00:21:16.978 Job: Nvme4n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:16.978 Job: Nvme4n1 ended in about 0.92 seconds with error 00:21:16.978 Verification LBA range: start 0x0 length 0x400 00:21:16.978 Nvme4n1 : 0.92 208.97 13.06 69.66 0.00 215488.45 14474.91 216097.84 00:21:16.978 Job: Nvme5n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:16.978 Job: Nvme5n1 ended in about 0.92 seconds with error 00:21:16.978 Verification LBA range: start 0x0 length 0x400 00:21:16.978 Nvme5n1 : 0.92 208.39 13.02 69.46 0.00 212167.68 28949.82 198773.54 00:21:16.978 Job: Nvme6n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:16.978 Job: Nvme6n1 ended in about 0.92 seconds with error 00:21:16.978 Verification LBA range: start 0x0 length 0x400 00:21:16.978 Nvme6n1 : 0.92 207.95 13.00 69.32 0.00 208723.26 17324.30 216097.84 00:21:16.978 Job: Nvme7n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:16.978 Job: Nvme7n1 ended in about 0.93 seconds with error 00:21:16.978 Verification LBA range: start 0x0 length 0x400 00:21:16.978 Nvme7n1 : 0.93 207.50 12.97 69.17 0.00 205241.43 15728.64 215186.03 00:21:16.978 Job: Nvme8n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:16.978 Job: Nvme8n1 ended in about 0.93 seconds with error 00:21:16.978 Verification LBA range: start 0x0 length 0x400 00:21:16.978 Nvme8n1 : 0.93 207.05 12.94 69.02 0.00 201781.20 16412.49 206067.98 00:21:16.978 Job: Nvme9n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:16.978 Verification LBA range: start 0x0 length 0x400 00:21:16.978 Nvme9n1 : 0.91 210.72 13.17 0.00 0.00 258438.90 18236.10 238892.97 00:21:16.978 Job: Nvme10n1 (Core Mask 0x1, workload: verify, depth: 64, IO size: 65536) 00:21:16.978 Job: Nvme10n1 ended in about 0.91 seconds with error 00:21:16.978 Verification LBA range: start 0x0 length 0x400 00:21:16.978 Nvme10n1 : 0.91 211.52 13.22 70.51 0.00 189058.45 15728.64 244363.80 00:21:16.978 =================================================================================================================== 00:21:16.978 Total : 2095.72 130.98 628.33 0.00 214286.77 13335.15 244363.80 00:21:16.978 [2024-07-15 18:07:10.645911] app.c:1053:spdk_app_stop: *WARNING*: spdk_app_stop'd on non-zero 00:21:16.978 [2024-07-15 18:07:10.645947] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode8] resetting controller 00:21:16.978 [2024-07-15 18:07:10.646246] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:16.978 [2024-07-15 18:07:10.646263] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x16f5050 with addr=10.0.0.2, port=4420 00:21:16.978 [2024-07-15 18:07:10.646273] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x16f5050 is same with the state(5) to be set 00:21:16.978 [2024-07-15 18:07:10.646461] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:16.978 [2024-07-15 18:07:10.646471] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x155d1d0 with addr=10.0.0.2, port=4420 00:21:16.978 [2024-07-15 18:07:10.646478] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x155d1d0 is same with the state(5) to be set 00:21:16.978 [2024-07-15 18:07:10.646716] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:16.978 [2024-07-15 18:07:10.646726] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x106f340 with addr=10.0.0.2, port=4420 00:21:16.978 [2024-07-15 18:07:10.646733] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x106f340 is same with the state(5) to be set 00:21:16.978 [2024-07-15 18:07:10.646919] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:16.978 [2024-07-15 18:07:10.646929] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1564b30 with addr=10.0.0.2, port=4420 00:21:16.978 [2024-07-15 18:07:10.646936] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1564b30 is same with the state(5) to be set 00:21:16.978 [2024-07-15 18:07:10.648318] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode2] resetting controller 00:21:16.979 [2024-07-15 18:07:10.648333] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:21:16.979 [2024-07-15 18:07:10.648342] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode10] resetting controller 00:21:16.979 [2024-07-15 18:07:10.648351] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode9] resetting controller 00:21:16.979 [2024-07-15 18:07:10.648673] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:16.979 [2024-07-15 18:07:10.648686] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1543190 with addr=10.0.0.2, port=4420 00:21:16.979 [2024-07-15 18:07:10.648694] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1543190 is same with the state(5) to be set 00:21:16.979 [2024-07-15 18:07:10.648947] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:16.979 [2024-07-15 18:07:10.648956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1567bf0 with addr=10.0.0.2, port=4420 00:21:16.979 [2024-07-15 18:07:10.648963] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1567bf0 is same with the state(5) to be set 00:21:16.979 [2024-07-15 18:07:10.648974] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x16f5050 (9): Bad file descriptor 00:21:16.979 [2024-07-15 18:07:10.648984] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x155d1d0 (9): Bad file descriptor 00:21:16.979 [2024-07-15 18:07:10.648993] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x106f340 (9): Bad file descriptor 00:21:16.979 [2024-07-15 18:07:10.649002] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1564b30 (9): Bad file descriptor 00:21:16.979 [2024-07-15 18:07:10.649033] bdev_nvme.c:2899:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:21:16.979 [2024-07-15 18:07:10.649043] bdev_nvme.c:2899:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:21:16.979 [2024-07-15 18:07:10.649053] bdev_nvme.c:2899:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:21:16.979 [2024-07-15 18:07:10.649063] bdev_nvme.c:2899:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:21:16.979 [2024-07-15 18:07:10.649330] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:16.979 [2024-07-15 18:07:10.649342] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x16ec8d0 with addr=10.0.0.2, port=4420 00:21:16.979 [2024-07-15 18:07:10.649349] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x16ec8d0 is same with the state(5) to be set 00:21:16.979 [2024-07-15 18:07:10.649528] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:16.979 [2024-07-15 18:07:10.649538] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1520c70 with addr=10.0.0.2, port=4420 00:21:16.979 [2024-07-15 18:07:10.649544] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1520c70 is same with the state(5) to be set 00:21:16.979 [2024-07-15 18:07:10.649674] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:16.979 [2024-07-15 18:07:10.649684] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x16d58b0 with addr=10.0.0.2, port=4420 00:21:16.979 [2024-07-15 18:07:10.649691] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x16d58b0 is same with the state(5) to be set 00:21:16.979 [2024-07-15 18:07:10.649854] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:21:16.979 [2024-07-15 18:07:10.649864] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x16ec0d0 with addr=10.0.0.2, port=4420 00:21:16.979 [2024-07-15 18:07:10.649870] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x16ec0d0 is same with the state(5) to be set 00:21:16.979 [2024-07-15 18:07:10.649879] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1543190 (9): Bad file descriptor 00:21:16.979 [2024-07-15 18:07:10.649887] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1567bf0 (9): Bad file descriptor 00:21:16.979 [2024-07-15 18:07:10.649895] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode3] Ctrlr is in error state 00:21:16.979 [2024-07-15 18:07:10.649901] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode3] controller reinitialization failed 00:21:16.979 [2024-07-15 18:07:10.649909] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode3] in failed state. 00:21:16.979 [2024-07-15 18:07:10.649919] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode4] Ctrlr is in error state 00:21:16.979 [2024-07-15 18:07:10.649925] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode4] controller reinitialization failed 00:21:16.979 [2024-07-15 18:07:10.649931] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode4] in failed state. 00:21:16.979 [2024-07-15 18:07:10.649942] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode5] Ctrlr is in error state 00:21:16.979 [2024-07-15 18:07:10.649948] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode5] controller reinitialization failed 00:21:16.979 [2024-07-15 18:07:10.649954] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode5] in failed state. 00:21:16.979 [2024-07-15 18:07:10.649963] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode6] Ctrlr is in error state 00:21:16.979 [2024-07-15 18:07:10.649969] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode6] controller reinitialization failed 00:21:16.979 [2024-07-15 18:07:10.649975] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode6] in failed state. 00:21:16.979 [2024-07-15 18:07:10.650040] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:16.979 [2024-07-15 18:07:10.650048] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:16.979 [2024-07-15 18:07:10.650054] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:16.979 [2024-07-15 18:07:10.650059] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:16.979 [2024-07-15 18:07:10.650070] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x16ec8d0 (9): Bad file descriptor 00:21:16.979 [2024-07-15 18:07:10.650078] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1520c70 (9): Bad file descriptor 00:21:16.979 [2024-07-15 18:07:10.650086] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x16d58b0 (9): Bad file descriptor 00:21:16.979 [2024-07-15 18:07:10.650094] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x16ec0d0 (9): Bad file descriptor 00:21:16.979 [2024-07-15 18:07:10.650101] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode7] Ctrlr is in error state 00:21:16.979 [2024-07-15 18:07:10.650107] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode7] controller reinitialization failed 00:21:16.979 [2024-07-15 18:07:10.650113] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode7] in failed state. 00:21:16.979 [2024-07-15 18:07:10.650121] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode8] Ctrlr is in error state 00:21:16.979 [2024-07-15 18:07:10.650127] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode8] controller reinitialization failed 00:21:16.979 [2024-07-15 18:07:10.650133] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode8] in failed state. 00:21:16.979 [2024-07-15 18:07:10.650157] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:16.979 [2024-07-15 18:07:10.650164] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:16.979 [2024-07-15 18:07:10.650169] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode2] Ctrlr is in error state 00:21:16.979 [2024-07-15 18:07:10.650175] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode2] controller reinitialization failed 00:21:16.979 [2024-07-15 18:07:10.650181] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode2] in failed state. 00:21:16.979 [2024-07-15 18:07:10.650189] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:21:16.979 [2024-07-15 18:07:10.650195] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:21:16.979 [2024-07-15 18:07:10.650201] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:21:16.979 [2024-07-15 18:07:10.650209] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode10] Ctrlr is in error state 00:21:16.979 [2024-07-15 18:07:10.650215] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode10] controller reinitialization failed 00:21:16.979 [2024-07-15 18:07:10.650221] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode10] in failed state. 00:21:16.979 [2024-07-15 18:07:10.650235] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode9] Ctrlr is in error state 00:21:16.979 [2024-07-15 18:07:10.650241] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode9] controller reinitialization failed 00:21:16.979 [2024-07-15 18:07:10.650248] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode9] in failed state. 00:21:16.979 [2024-07-15 18:07:10.650274] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:16.979 [2024-07-15 18:07:10.650281] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:16.979 [2024-07-15 18:07:10.650287] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:16.979 [2024-07-15 18:07:10.650293] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:21:17.544 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@136 -- # nvmfpid= 00:21:17.544 18:07:10 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@139 -- # sleep 1 00:21:18.481 18:07:11 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@142 -- # kill -9 652222 00:21:18.481 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/shutdown.sh: line 142: kill: (652222) - No such process 00:21:18.481 18:07:11 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@142 -- # true 00:21:18.481 18:07:11 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@144 -- # stoptarget 00:21:18.481 18:07:11 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@41 -- # rm -f ./local-job0-0-verify.state 00:21:18.481 18:07:11 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@42 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/bdevperf.conf 00:21:18.481 18:07:11 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@43 -- # rm -rf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/rpcs.txt 00:21:18.481 18:07:12 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- target/shutdown.sh@45 -- # nvmftestfini 00:21:18.481 18:07:12 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@488 -- # nvmfcleanup 00:21:18.481 18:07:12 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@117 -- # sync 00:21:18.481 18:07:12 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:21:18.481 18:07:12 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@120 -- # set +e 00:21:18.481 18:07:12 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@121 -- # for i in {1..20} 00:21:18.481 18:07:12 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:21:18.481 rmmod nvme_tcp 00:21:18.481 rmmod nvme_fabrics 00:21:18.481 rmmod nvme_keyring 00:21:18.481 18:07:12 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:21:18.481 18:07:12 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@124 -- # set -e 00:21:18.481 18:07:12 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@125 -- # return 0 00:21:18.481 18:07:12 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@489 -- # '[' -n '' ']' 00:21:18.481 18:07:12 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:21:18.481 18:07:12 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:21:18.481 18:07:12 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:21:18.481 18:07:12 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:21:18.481 18:07:12 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@278 -- # remove_spdk_ns 00:21:18.481 18:07:12 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:18.481 18:07:12 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:18.481 18:07:12 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:20.428 18:07:14 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:21:20.428 00:21:20.428 real 0m7.912s 00:21:20.428 user 0m19.825s 00:21:20.428 sys 0m1.307s 00:21:20.428 18:07:14 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@1124 -- # xtrace_disable 00:21:20.428 18:07:14 nvmf_tcp.nvmf_shutdown.nvmf_shutdown_tc3 -- common/autotest_common.sh@10 -- # set +x 00:21:20.428 ************************************ 00:21:20.428 END TEST nvmf_shutdown_tc3 00:21:20.428 ************************************ 00:21:20.687 18:07:14 nvmf_tcp.nvmf_shutdown -- common/autotest_common.sh@1142 -- # return 0 00:21:20.687 18:07:14 nvmf_tcp.nvmf_shutdown -- target/shutdown.sh@151 -- # trap - SIGINT SIGTERM EXIT 00:21:20.687 00:21:20.687 real 0m31.113s 00:21:20.687 user 1m20.427s 00:21:20.687 sys 0m7.882s 00:21:20.687 18:07:14 nvmf_tcp.nvmf_shutdown -- common/autotest_common.sh@1124 -- # xtrace_disable 00:21:20.687 18:07:14 nvmf_tcp.nvmf_shutdown -- common/autotest_common.sh@10 -- # set +x 00:21:20.687 ************************************ 00:21:20.687 END TEST nvmf_shutdown 00:21:20.687 ************************************ 00:21:20.687 18:07:14 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:21:20.687 18:07:14 nvmf_tcp -- nvmf/nvmf.sh@86 -- # timing_exit target 00:21:20.687 18:07:14 nvmf_tcp -- common/autotest_common.sh@728 -- # xtrace_disable 00:21:20.687 18:07:14 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:21:20.687 18:07:14 nvmf_tcp -- nvmf/nvmf.sh@88 -- # timing_enter host 00:21:20.687 18:07:14 nvmf_tcp -- common/autotest_common.sh@722 -- # xtrace_disable 00:21:20.687 18:07:14 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:21:20.687 18:07:14 nvmf_tcp -- nvmf/nvmf.sh@90 -- # [[ 0 -eq 0 ]] 00:21:20.687 18:07:14 nvmf_tcp -- nvmf/nvmf.sh@91 -- # run_test nvmf_multicontroller /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multicontroller.sh --transport=tcp 00:21:20.687 18:07:14 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:21:20.687 18:07:14 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:21:20.687 18:07:14 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:21:20.687 ************************************ 00:21:20.687 START TEST nvmf_multicontroller 00:21:20.687 ************************************ 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multicontroller.sh --transport=tcp 00:21:20.687 * Looking for test storage... 00:21:20.687 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@7 -- # uname -s 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- paths/export.sh@5 -- # export PATH 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@47 -- # : 0 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@51 -- # have_pci_nics=0 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@11 -- # MALLOC_BDEV_SIZE=64 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@13 -- # NVMF_HOST_FIRST_PORT=60000 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@14 -- # NVMF_HOST_SECOND_PORT=60001 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@16 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@18 -- # '[' tcp == rdma ']' 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@23 -- # nvmftestinit 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@448 -- # prepare_net_devs 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@410 -- # local -g is_hw=no 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@412 -- # remove_spdk_ns 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@285 -- # xtrace_disable 00:21:20.687 18:07:14 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@291 -- # pci_devs=() 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@291 -- # local -a pci_devs 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@292 -- # pci_net_devs=() 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@293 -- # pci_drivers=() 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@293 -- # local -A pci_drivers 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@295 -- # net_devs=() 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@295 -- # local -ga net_devs 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@296 -- # e810=() 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@296 -- # local -ga e810 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@297 -- # x722=() 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@297 -- # local -ga x722 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@298 -- # mlx=() 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@298 -- # local -ga mlx 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:21:25.961 Found 0000:86:00.0 (0x8086 - 0x159b) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:21:25.961 Found 0000:86:00.1 (0x8086 - 0x159b) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@390 -- # [[ up == up ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:21:25.961 Found net devices under 0000:86:00.0: cvl_0_0 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@390 -- # [[ up == up ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:21:25.961 Found net devices under 0000:86:00.1: cvl_0_1 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@414 -- # is_hw=yes 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:21:25.961 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:21:25.962 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:21:25.962 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:21:25.962 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:21:25.962 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:21:25.962 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:21:25.962 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:21:25.962 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:21:25.962 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:21:25.962 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:21:25.962 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:21:25.962 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:21:25.962 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:21:25.962 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:21:25.962 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:21:26.220 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:21:26.220 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:21:26.220 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.202 ms 00:21:26.220 00:21:26.220 --- 10.0.0.2 ping statistics --- 00:21:26.220 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:26.220 rtt min/avg/max/mdev = 0.202/0.202/0.202/0.000 ms 00:21:26.220 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:21:26.220 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:21:26.220 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.229 ms 00:21:26.220 00:21:26.220 --- 10.0.0.1 ping statistics --- 00:21:26.220 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:26.220 rtt min/avg/max/mdev = 0.229/0.229/0.229/0.000 ms 00:21:26.220 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@422 -- # return 0 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@25 -- # nvmfappstart -m 0xE 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@722 -- # xtrace_disable 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@481 -- # nvmfpid=656496 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@482 -- # waitforlisten 656496 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@829 -- # '[' -z 656496 ']' 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@834 -- # local max_retries=100 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:26.221 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@838 -- # xtrace_disable 00:21:26.221 18:07:19 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:26.221 [2024-07-15 18:07:19.787940] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:21:26.221 [2024-07-15 18:07:19.787990] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:26.221 EAL: No free 2048 kB hugepages reported on node 1 00:21:26.221 [2024-07-15 18:07:19.845901] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:21:26.221 [2024-07-15 18:07:19.925630] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:26.221 [2024-07-15 18:07:19.925664] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:26.221 [2024-07-15 18:07:19.925671] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:26.221 [2024-07-15 18:07:19.925677] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:26.221 [2024-07-15 18:07:19.925682] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:26.221 [2024-07-15 18:07:19.925780] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:21:26.221 [2024-07-15 18:07:19.925872] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:21:26.221 [2024-07-15 18:07:19.925873] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@862 -- # return 0 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@728 -- # xtrace_disable 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@27 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:27.157 [2024-07-15 18:07:20.645364] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@29 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:27.157 Malloc0 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@30 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@31 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@33 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:27.157 [2024-07-15 18:07:20.706053] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@34 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:27.157 [2024-07-15 18:07:20.714002] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@36 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc1 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:27.157 Malloc1 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@37 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 -a -s SPDK00000000000002 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@38 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 Malloc1 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@40 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@41 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4421 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@44 -- # bdevperf_pid=656657 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w write -t 1 -f 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@46 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; pap "$testdir/try.txt"; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@47 -- # waitforlisten 656657 /var/tmp/bdevperf.sock 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@829 -- # '[' -z 656657 ']' 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@834 -- # local max_retries=100 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:21:27.157 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@838 -- # xtrace_disable 00:21:27.157 18:07:20 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:28.091 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:21:28.091 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@862 -- # return 0 00:21:28.091 18:07:21 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@50 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 00:21:28.091 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:28.091 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:28.349 NVMe0n1 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@54 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@54 -- # grep -c NVMe 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:28.349 1 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@60 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 -q nqn.2021-09-7.io.spdk:00001 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@648 -- # local es=0 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@650 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 -q nqn.2021-09-7.io.spdk:00001 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@636 -- # local arg=rpc_cmd 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@640 -- # type -t rpc_cmd 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@651 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 -q nqn.2021-09-7.io.spdk:00001 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:28.349 request: 00:21:28.349 { 00:21:28.349 "name": "NVMe0", 00:21:28.349 "trtype": "tcp", 00:21:28.349 "traddr": "10.0.0.2", 00:21:28.349 "adrfam": "ipv4", 00:21:28.349 "trsvcid": "4420", 00:21:28.349 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:21:28.349 "hostnqn": "nqn.2021-09-7.io.spdk:00001", 00:21:28.349 "hostaddr": "10.0.0.2", 00:21:28.349 "hostsvcid": "60000", 00:21:28.349 "prchk_reftag": false, 00:21:28.349 "prchk_guard": false, 00:21:28.349 "hdgst": false, 00:21:28.349 "ddgst": false, 00:21:28.349 "method": "bdev_nvme_attach_controller", 00:21:28.349 "req_id": 1 00:21:28.349 } 00:21:28.349 Got JSON-RPC error response 00:21:28.349 response: 00:21:28.349 { 00:21:28.349 "code": -114, 00:21:28.349 "message": "A controller named NVMe0 already exists with the specified network path\n" 00:21:28.349 } 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 1 == 0 ]] 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@651 -- # es=1 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@65 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -i 10.0.0.2 -c 60000 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@648 -- # local es=0 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@650 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -i 10.0.0.2 -c 60000 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@636 -- # local arg=rpc_cmd 00:21:28.349 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@640 -- # type -t rpc_cmd 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@651 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode2 -i 10.0.0.2 -c 60000 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:28.350 request: 00:21:28.350 { 00:21:28.350 "name": "NVMe0", 00:21:28.350 "trtype": "tcp", 00:21:28.350 "traddr": "10.0.0.2", 00:21:28.350 "adrfam": "ipv4", 00:21:28.350 "trsvcid": "4420", 00:21:28.350 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:21:28.350 "hostaddr": "10.0.0.2", 00:21:28.350 "hostsvcid": "60000", 00:21:28.350 "prchk_reftag": false, 00:21:28.350 "prchk_guard": false, 00:21:28.350 "hdgst": false, 00:21:28.350 "ddgst": false, 00:21:28.350 "method": "bdev_nvme_attach_controller", 00:21:28.350 "req_id": 1 00:21:28.350 } 00:21:28.350 Got JSON-RPC error response 00:21:28.350 response: 00:21:28.350 { 00:21:28.350 "code": -114, 00:21:28.350 "message": "A controller named NVMe0 already exists with the specified network path\n" 00:21:28.350 } 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 1 == 0 ]] 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@651 -- # es=1 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@69 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 -x disable 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@648 -- # local es=0 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@650 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 -x disable 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@636 -- # local arg=rpc_cmd 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@640 -- # type -t rpc_cmd 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@651 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 -x disable 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:28.350 request: 00:21:28.350 { 00:21:28.350 "name": "NVMe0", 00:21:28.350 "trtype": "tcp", 00:21:28.350 "traddr": "10.0.0.2", 00:21:28.350 "adrfam": "ipv4", 00:21:28.350 "trsvcid": "4420", 00:21:28.350 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:21:28.350 "hostaddr": "10.0.0.2", 00:21:28.350 "hostsvcid": "60000", 00:21:28.350 "prchk_reftag": false, 00:21:28.350 "prchk_guard": false, 00:21:28.350 "hdgst": false, 00:21:28.350 "ddgst": false, 00:21:28.350 "multipath": "disable", 00:21:28.350 "method": "bdev_nvme_attach_controller", 00:21:28.350 "req_id": 1 00:21:28.350 } 00:21:28.350 Got JSON-RPC error response 00:21:28.350 response: 00:21:28.350 { 00:21:28.350 "code": -114, 00:21:28.350 "message": "A controller named NVMe0 already exists and multipath is disabled\n" 00:21:28.350 } 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 1 == 0 ]] 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@651 -- # es=1 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@74 -- # NOT rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 -x failover 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@648 -- # local es=0 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@650 -- # valid_exec_arg rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 -x failover 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@636 -- # local arg=rpc_cmd 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@640 -- # type -t rpc_cmd 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@651 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 -x failover 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:28.350 request: 00:21:28.350 { 00:21:28.350 "name": "NVMe0", 00:21:28.350 "trtype": "tcp", 00:21:28.350 "traddr": "10.0.0.2", 00:21:28.350 "adrfam": "ipv4", 00:21:28.350 "trsvcid": "4420", 00:21:28.350 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:21:28.350 "hostaddr": "10.0.0.2", 00:21:28.350 "hostsvcid": "60000", 00:21:28.350 "prchk_reftag": false, 00:21:28.350 "prchk_guard": false, 00:21:28.350 "hdgst": false, 00:21:28.350 "ddgst": false, 00:21:28.350 "multipath": "failover", 00:21:28.350 "method": "bdev_nvme_attach_controller", 00:21:28.350 "req_id": 1 00:21:28.350 } 00:21:28.350 Got JSON-RPC error response 00:21:28.350 response: 00:21:28.350 { 00:21:28.350 "code": -114, 00:21:28.350 "message": "A controller named NVMe0 already exists with the specified network path\n" 00:21:28.350 } 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 1 == 0 ]] 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@651 -- # es=1 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@79 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:28.350 18:07:21 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:28.350 00:21:28.350 18:07:22 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:28.350 18:07:22 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@83 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:21:28.350 18:07:22 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:28.350 18:07:22 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:28.350 18:07:22 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:28.350 18:07:22 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@87 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe1 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -i 10.0.0.2 -c 60000 00:21:28.350 18:07:22 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:28.350 18:07:22 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:28.608 00:21:28.608 18:07:22 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:28.608 18:07:22 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@90 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:21:28.608 18:07:22 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@90 -- # grep -c NVMe 00:21:28.608 18:07:22 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:28.608 18:07:22 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:28.608 18:07:22 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:28.608 18:07:22 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@90 -- # '[' 2 '!=' 2 ']' 00:21:28.608 18:07:22 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@95 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:21:29.983 0 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@98 -- # rpc_cmd -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe1 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@100 -- # killprocess 656657 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@948 -- # '[' -z 656657 ']' 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@952 -- # kill -0 656657 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@953 -- # uname 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 656657 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@966 -- # echo 'killing process with pid 656657' 00:21:29.983 killing process with pid 656657 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@967 -- # kill 656657 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@972 -- # wait 656657 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@102 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@103 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode2 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@105 -- # trap - SIGINT SIGTERM EXIT 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@107 -- # pap /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@1612 -- # read -r file 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@1611 -- # find /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt -type f 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@1611 -- # sort -u 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@1613 -- # cat 00:21:29.983 --- /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt --- 00:21:29.983 [2024-07-15 18:07:20.816798] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:21:29.983 [2024-07-15 18:07:20.816849] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid656657 ] 00:21:29.983 EAL: No free 2048 kB hugepages reported on node 1 00:21:29.983 [2024-07-15 18:07:20.871347] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:29.983 [2024-07-15 18:07:20.952361] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:21:29.983 [2024-07-15 18:07:22.185425] bdev.c:4613:bdev_name_add: *ERROR*: Bdev name b94d3f1c-2162-4d99-bf7e-0b9b03a789a0 already exists 00:21:29.983 [2024-07-15 18:07:22.185455] bdev.c:7722:bdev_register: *ERROR*: Unable to add uuid:b94d3f1c-2162-4d99-bf7e-0b9b03a789a0 alias for bdev NVMe1n1 00:21:29.983 [2024-07-15 18:07:22.185463] bdev_nvme.c:4317:nvme_bdev_create: *ERROR*: spdk_bdev_register() failed 00:21:29.983 Running I/O for 1 seconds... 00:21:29.983 00:21:29.983 Latency(us) 00:21:29.983 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:21:29.983 Job: NVMe0n1 (Core Mask 0x1, workload: write, depth: 128, IO size: 4096) 00:21:29.983 NVMe0n1 : 1.01 24615.92 96.16 0.00 0.00 5188.91 3177.07 11682.50 00:21:29.983 =================================================================================================================== 00:21:29.983 Total : 24615.92 96.16 0.00 0.00 5188.91 3177.07 11682.50 00:21:29.983 Received shutdown signal, test time was about 1.000000 seconds 00:21:29.983 00:21:29.983 Latency(us) 00:21:29.983 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:21:29.983 =================================================================================================================== 00:21:29.983 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:21:29.983 --- /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt --- 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@1618 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@1612 -- # read -r file 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- host/multicontroller.sh@108 -- # nvmftestfini 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@488 -- # nvmfcleanup 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@117 -- # sync 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@120 -- # set +e 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@121 -- # for i in {1..20} 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:21:29.983 rmmod nvme_tcp 00:21:29.983 rmmod nvme_fabrics 00:21:29.983 rmmod nvme_keyring 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@124 -- # set -e 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@125 -- # return 0 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@489 -- # '[' -n 656496 ']' 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@490 -- # killprocess 656496 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@948 -- # '[' -z 656496 ']' 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@952 -- # kill -0 656496 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@953 -- # uname 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:21:29.983 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 656496 00:21:30.242 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:21:30.242 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:21:30.242 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@966 -- # echo 'killing process with pid 656496' 00:21:30.242 killing process with pid 656496 00:21:30.242 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@967 -- # kill 656496 00:21:30.242 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@972 -- # wait 656496 00:21:30.242 18:07:23 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:21:30.242 18:07:23 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:21:30.242 18:07:23 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:21:30.242 18:07:23 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:21:30.242 18:07:23 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@278 -- # remove_spdk_ns 00:21:30.242 18:07:23 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:30.242 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:30.242 18:07:23 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:32.777 18:07:25 nvmf_tcp.nvmf_multicontroller -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:21:32.777 00:21:32.777 real 0m11.713s 00:21:32.777 user 0m16.642s 00:21:32.777 sys 0m4.726s 00:21:32.777 18:07:26 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@1124 -- # xtrace_disable 00:21:32.777 18:07:26 nvmf_tcp.nvmf_multicontroller -- common/autotest_common.sh@10 -- # set +x 00:21:32.777 ************************************ 00:21:32.777 END TEST nvmf_multicontroller 00:21:32.777 ************************************ 00:21:32.777 18:07:26 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:21:32.777 18:07:26 nvmf_tcp -- nvmf/nvmf.sh@92 -- # run_test nvmf_aer /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/aer.sh --transport=tcp 00:21:32.777 18:07:26 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:21:32.777 18:07:26 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:21:32.777 18:07:26 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:21:32.777 ************************************ 00:21:32.777 START TEST nvmf_aer 00:21:32.777 ************************************ 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/aer.sh --transport=tcp 00:21:32.777 * Looking for test storage... 00:21:32.777 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- host/aer.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@7 -- # uname -s 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:21:32.777 18:07:26 nvmf_tcp.nvmf_aer -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- paths/export.sh@5 -- # export PATH 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@47 -- # : 0 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@51 -- # have_pci_nics=0 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- host/aer.sh@11 -- # nvmftestinit 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@448 -- # prepare_net_devs 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@410 -- # local -g is_hw=no 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@412 -- # remove_spdk_ns 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- nvmf/common.sh@285 -- # xtrace_disable 00:21:32.778 18:07:26 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@291 -- # pci_devs=() 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@291 -- # local -a pci_devs 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@292 -- # pci_net_devs=() 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@293 -- # pci_drivers=() 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@293 -- # local -A pci_drivers 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@295 -- # net_devs=() 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@295 -- # local -ga net_devs 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@296 -- # e810=() 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@296 -- # local -ga e810 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@297 -- # x722=() 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@297 -- # local -ga x722 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@298 -- # mlx=() 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@298 -- # local -ga mlx 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:21:38.047 Found 0000:86:00.0 (0x8086 - 0x159b) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:21:38.047 Found 0000:86:00.1 (0x8086 - 0x159b) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@390 -- # [[ up == up ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:21:38.047 Found net devices under 0000:86:00.0: cvl_0_0 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@390 -- # [[ up == up ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:21:38.047 Found net devices under 0000:86:00.1: cvl_0_1 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@414 -- # is_hw=yes 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:21:38.047 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:21:38.047 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.311 ms 00:21:38.047 00:21:38.047 --- 10.0.0.2 ping statistics --- 00:21:38.047 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:38.047 rtt min/avg/max/mdev = 0.311/0.311/0.311/0.000 ms 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:21:38.047 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:21:38.047 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.255 ms 00:21:38.047 00:21:38.047 --- 10.0.0.1 ping statistics --- 00:21:38.047 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:38.047 rtt min/avg/max/mdev = 0.255/0.255/0.255/0.000 ms 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@422 -- # return 0 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:21:38.047 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:21:38.048 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:21:38.048 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:21:38.048 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:21:38.048 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:21:38.048 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:21:38.048 18:07:31 nvmf_tcp.nvmf_aer -- host/aer.sh@12 -- # nvmfappstart -m 0xF 00:21:38.048 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:21:38.048 18:07:31 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@722 -- # xtrace_disable 00:21:38.048 18:07:31 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:21:38.048 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@481 -- # nvmfpid=660521 00:21:38.048 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@482 -- # waitforlisten 660521 00:21:38.048 18:07:31 nvmf_tcp.nvmf_aer -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:21:38.048 18:07:31 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@829 -- # '[' -z 660521 ']' 00:21:38.048 18:07:31 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:38.048 18:07:31 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@834 -- # local max_retries=100 00:21:38.048 18:07:31 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:38.048 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:38.048 18:07:31 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@838 -- # xtrace_disable 00:21:38.048 18:07:31 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:21:38.048 [2024-07-15 18:07:31.481461] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:21:38.048 [2024-07-15 18:07:31.481509] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:38.048 EAL: No free 2048 kB hugepages reported on node 1 00:21:38.048 [2024-07-15 18:07:31.538733] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:21:38.048 [2024-07-15 18:07:31.619556] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:38.048 [2024-07-15 18:07:31.619590] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:38.048 [2024-07-15 18:07:31.619598] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:38.048 [2024-07-15 18:07:31.619604] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:38.048 [2024-07-15 18:07:31.619610] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:38.048 [2024-07-15 18:07:31.619643] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:21:38.048 [2024-07-15 18:07:31.619741] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:21:38.048 [2024-07-15 18:07:31.619818] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:21:38.048 [2024-07-15 18:07:31.619819] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:21:38.613 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:21:38.613 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@862 -- # return 0 00:21:38.613 18:07:32 nvmf_tcp.nvmf_aer -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:21:38.613 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@728 -- # xtrace_disable 00:21:38.613 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- host/aer.sh@14 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:21:38.871 [2024-07-15 18:07:32.349151] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- host/aer.sh@16 -- # rpc_cmd bdev_malloc_create 64 512 --name Malloc0 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:21:38.871 Malloc0 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- host/aer.sh@17 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 2 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- host/aer.sh@18 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- host/aer.sh@19 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:21:38.871 [2024-07-15 18:07:32.400967] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- host/aer.sh@21 -- # rpc_cmd nvmf_get_subsystems 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:38.871 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:21:38.871 [ 00:21:38.871 { 00:21:38.871 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:21:38.871 "subtype": "Discovery", 00:21:38.871 "listen_addresses": [], 00:21:38.871 "allow_any_host": true, 00:21:38.871 "hosts": [] 00:21:38.871 }, 00:21:38.871 { 00:21:38.871 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:21:38.871 "subtype": "NVMe", 00:21:38.871 "listen_addresses": [ 00:21:38.871 { 00:21:38.871 "trtype": "TCP", 00:21:38.872 "adrfam": "IPv4", 00:21:38.872 "traddr": "10.0.0.2", 00:21:38.872 "trsvcid": "4420" 00:21:38.872 } 00:21:38.872 ], 00:21:38.872 "allow_any_host": true, 00:21:38.872 "hosts": [], 00:21:38.872 "serial_number": "SPDK00000000000001", 00:21:38.872 "model_number": "SPDK bdev Controller", 00:21:38.872 "max_namespaces": 2, 00:21:38.872 "min_cntlid": 1, 00:21:38.872 "max_cntlid": 65519, 00:21:38.872 "namespaces": [ 00:21:38.872 { 00:21:38.872 "nsid": 1, 00:21:38.872 "bdev_name": "Malloc0", 00:21:38.872 "name": "Malloc0", 00:21:38.872 "nguid": "06F45E3332A940DB94F74F6D70F0B188", 00:21:38.872 "uuid": "06f45e33-32a9-40db-94f7-4f6d70f0b188" 00:21:38.872 } 00:21:38.872 ] 00:21:38.872 } 00:21:38.872 ] 00:21:38.872 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:38.872 18:07:32 nvmf_tcp.nvmf_aer -- host/aer.sh@23 -- # AER_TOUCH_FILE=/tmp/aer_touch_file 00:21:38.872 18:07:32 nvmf_tcp.nvmf_aer -- host/aer.sh@24 -- # rm -f /tmp/aer_touch_file 00:21:38.872 18:07:32 nvmf_tcp.nvmf_aer -- host/aer.sh@33 -- # aerpid=660763 00:21:38.872 18:07:32 nvmf_tcp.nvmf_aer -- host/aer.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvme/aer/aer -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' -n 2 -t /tmp/aer_touch_file 00:21:38.872 18:07:32 nvmf_tcp.nvmf_aer -- host/aer.sh@36 -- # waitforfile /tmp/aer_touch_file 00:21:38.872 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@1265 -- # local i=0 00:21:38.872 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@1266 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:21:38.872 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@1267 -- # '[' 0 -lt 200 ']' 00:21:38.872 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@1268 -- # i=1 00:21:38.872 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@1269 -- # sleep 0.1 00:21:38.872 EAL: No free 2048 kB hugepages reported on node 1 00:21:38.872 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@1266 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:21:38.872 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@1267 -- # '[' 1 -lt 200 ']' 00:21:38.872 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@1268 -- # i=2 00:21:38.872 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@1269 -- # sleep 0.1 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@1266 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@1267 -- # '[' 2 -lt 200 ']' 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@1268 -- # i=3 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@1269 -- # sleep 0.1 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@1266 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@1272 -- # '[' '!' -e /tmp/aer_touch_file ']' 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@1276 -- # return 0 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- host/aer.sh@39 -- # rpc_cmd bdev_malloc_create 64 4096 --name Malloc1 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:21:39.130 Malloc1 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- host/aer.sh@40 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 -n 2 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- host/aer.sh@41 -- # rpc_cmd nvmf_get_subsystems 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:21:39.130 Asynchronous Event Request test 00:21:39.130 Attaching to 10.0.0.2 00:21:39.130 Attached to 10.0.0.2 00:21:39.130 Registering asynchronous event callbacks... 00:21:39.130 Starting namespace attribute notice tests for all controllers... 00:21:39.130 10.0.0.2: aer_cb for log page 4, aen_event_type: 0x02, aen_event_info: 0x00 00:21:39.130 aer_cb - Changed Namespace 00:21:39.130 Cleaning up... 00:21:39.130 [ 00:21:39.130 { 00:21:39.130 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:21:39.130 "subtype": "Discovery", 00:21:39.130 "listen_addresses": [], 00:21:39.130 "allow_any_host": true, 00:21:39.130 "hosts": [] 00:21:39.130 }, 00:21:39.130 { 00:21:39.130 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:21:39.130 "subtype": "NVMe", 00:21:39.130 "listen_addresses": [ 00:21:39.130 { 00:21:39.130 "trtype": "TCP", 00:21:39.130 "adrfam": "IPv4", 00:21:39.130 "traddr": "10.0.0.2", 00:21:39.130 "trsvcid": "4420" 00:21:39.130 } 00:21:39.130 ], 00:21:39.130 "allow_any_host": true, 00:21:39.130 "hosts": [], 00:21:39.130 "serial_number": "SPDK00000000000001", 00:21:39.130 "model_number": "SPDK bdev Controller", 00:21:39.130 "max_namespaces": 2, 00:21:39.130 "min_cntlid": 1, 00:21:39.130 "max_cntlid": 65519, 00:21:39.130 "namespaces": [ 00:21:39.130 { 00:21:39.130 "nsid": 1, 00:21:39.130 "bdev_name": "Malloc0", 00:21:39.130 "name": "Malloc0", 00:21:39.130 "nguid": "06F45E3332A940DB94F74F6D70F0B188", 00:21:39.130 "uuid": "06f45e33-32a9-40db-94f7-4f6d70f0b188" 00:21:39.130 }, 00:21:39.130 { 00:21:39.130 "nsid": 2, 00:21:39.130 "bdev_name": "Malloc1", 00:21:39.130 "name": "Malloc1", 00:21:39.130 "nguid": "29BF4DF9331C49C1B039BAEB8CF2AE5A", 00:21:39.130 "uuid": "29bf4df9-331c-49c1-b039-baeb8cf2ae5a" 00:21:39.130 } 00:21:39.130 ] 00:21:39.130 } 00:21:39.130 ] 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- host/aer.sh@43 -- # wait 660763 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- host/aer.sh@45 -- # rpc_cmd bdev_malloc_delete Malloc0 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- host/aer.sh@46 -- # rpc_cmd bdev_malloc_delete Malloc1 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- host/aer.sh@47 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:39.130 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- host/aer.sh@49 -- # trap - SIGINT SIGTERM EXIT 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- host/aer.sh@51 -- # nvmftestfini 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- nvmf/common.sh@488 -- # nvmfcleanup 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- nvmf/common.sh@117 -- # sync 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- nvmf/common.sh@120 -- # set +e 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- nvmf/common.sh@121 -- # for i in {1..20} 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:21:39.389 rmmod nvme_tcp 00:21:39.389 rmmod nvme_fabrics 00:21:39.389 rmmod nvme_keyring 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- nvmf/common.sh@124 -- # set -e 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- nvmf/common.sh@125 -- # return 0 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- nvmf/common.sh@489 -- # '[' -n 660521 ']' 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- nvmf/common.sh@490 -- # killprocess 660521 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@948 -- # '[' -z 660521 ']' 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@952 -- # kill -0 660521 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@953 -- # uname 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 660521 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@966 -- # echo 'killing process with pid 660521' 00:21:39.389 killing process with pid 660521 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@967 -- # kill 660521 00:21:39.389 18:07:32 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@972 -- # wait 660521 00:21:39.659 18:07:33 nvmf_tcp.nvmf_aer -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:21:39.659 18:07:33 nvmf_tcp.nvmf_aer -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:21:39.659 18:07:33 nvmf_tcp.nvmf_aer -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:21:39.659 18:07:33 nvmf_tcp.nvmf_aer -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:21:39.659 18:07:33 nvmf_tcp.nvmf_aer -- nvmf/common.sh@278 -- # remove_spdk_ns 00:21:39.659 18:07:33 nvmf_tcp.nvmf_aer -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:39.659 18:07:33 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:39.659 18:07:33 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:41.669 18:07:35 nvmf_tcp.nvmf_aer -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:21:41.669 00:21:41.669 real 0m9.127s 00:21:41.669 user 0m7.598s 00:21:41.669 sys 0m4.383s 00:21:41.669 18:07:35 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@1124 -- # xtrace_disable 00:21:41.669 18:07:35 nvmf_tcp.nvmf_aer -- common/autotest_common.sh@10 -- # set +x 00:21:41.669 ************************************ 00:21:41.669 END TEST nvmf_aer 00:21:41.669 ************************************ 00:21:41.669 18:07:35 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:21:41.669 18:07:35 nvmf_tcp -- nvmf/nvmf.sh@93 -- # run_test nvmf_async_init /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/async_init.sh --transport=tcp 00:21:41.669 18:07:35 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:21:41.669 18:07:35 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:21:41.669 18:07:35 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:21:41.669 ************************************ 00:21:41.669 START TEST nvmf_async_init 00:21:41.669 ************************************ 00:21:41.669 18:07:35 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/async_init.sh --transport=tcp 00:21:41.669 * Looking for test storage... 00:21:41.669 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:21:41.669 18:07:35 nvmf_tcp.nvmf_async_init -- host/async_init.sh@11 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:21:41.669 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@7 -- # uname -s 00:21:41.669 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:21:41.669 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:21:41.669 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:21:41.669 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:21:41.669 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:21:41.669 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:21:41.669 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:21:41.669 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:21:41.669 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:21:41.669 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:21:41.669 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:21:41.669 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:21:41.669 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- paths/export.sh@5 -- # export PATH 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@47 -- # : 0 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@51 -- # have_pci_nics=0 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- host/async_init.sh@13 -- # null_bdev_size=1024 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- host/async_init.sh@14 -- # null_block_size=512 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- host/async_init.sh@15 -- # null_bdev=null0 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- host/async_init.sh@16 -- # nvme_bdev=nvme0 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- host/async_init.sh@20 -- # uuidgen 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- host/async_init.sh@20 -- # tr -d - 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- host/async_init.sh@20 -- # nguid=16a3735b782d4338a53f5f5fc9d56ad1 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- host/async_init.sh@22 -- # nvmftestinit 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@448 -- # prepare_net_devs 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@410 -- # local -g is_hw=no 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@412 -- # remove_spdk_ns 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@285 -- # xtrace_disable 00:21:41.670 18:07:35 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@291 -- # pci_devs=() 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@291 -- # local -a pci_devs 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@292 -- # pci_net_devs=() 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@293 -- # pci_drivers=() 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@293 -- # local -A pci_drivers 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@295 -- # net_devs=() 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@295 -- # local -ga net_devs 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@296 -- # e810=() 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@296 -- # local -ga e810 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@297 -- # x722=() 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@297 -- # local -ga x722 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@298 -- # mlx=() 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@298 -- # local -ga mlx 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:21:46.938 Found 0000:86:00.0 (0x8086 - 0x159b) 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:46.938 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:21:46.938 Found 0000:86:00.1 (0x8086 - 0x159b) 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@390 -- # [[ up == up ]] 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:21:46.939 Found net devices under 0000:86:00.0: cvl_0_0 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@390 -- # [[ up == up ]] 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:21:46.939 Found net devices under 0000:86:00.1: cvl_0_1 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@414 -- # is_hw=yes 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:21:46.939 18:07:39 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:21:46.939 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:21:46.939 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.261 ms 00:21:46.939 00:21:46.939 --- 10.0.0.2 ping statistics --- 00:21:46.939 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:46.939 rtt min/avg/max/mdev = 0.261/0.261/0.261/0.000 ms 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:21:46.939 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:21:46.939 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.243 ms 00:21:46.939 00:21:46.939 --- 10.0.0.1 ping statistics --- 00:21:46.939 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:46.939 rtt min/avg/max/mdev = 0.243/0.243/0.243/0.000 ms 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@422 -- # return 0 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- host/async_init.sh@23 -- # nvmfappstart -m 0x1 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@722 -- # xtrace_disable 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@481 -- # nvmfpid=664066 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@482 -- # waitforlisten 664066 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@829 -- # '[' -z 664066 ']' 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@834 -- # local max_retries=100 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:46.939 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x1 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@838 -- # xtrace_disable 00:21:46.939 18:07:40 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:46.939 [2024-07-15 18:07:40.185319] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:21:46.939 [2024-07-15 18:07:40.185361] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:46.939 EAL: No free 2048 kB hugepages reported on node 1 00:21:46.939 [2024-07-15 18:07:40.240819] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:21:46.939 [2024-07-15 18:07:40.320820] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:46.939 [2024-07-15 18:07:40.320854] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:46.939 [2024-07-15 18:07:40.320862] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:46.939 [2024-07-15 18:07:40.320868] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:46.939 [2024-07-15 18:07:40.320873] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:46.939 [2024-07-15 18:07:40.320896] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:21:47.507 18:07:40 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:21:47.507 18:07:40 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@862 -- # return 0 00:21:47.507 18:07:40 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:21:47.507 18:07:40 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@728 -- # xtrace_disable 00:21:47.507 18:07:40 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@26 -- # rpc_cmd nvmf_create_transport -t tcp -o 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:47.507 [2024-07-15 18:07:41.028980] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@27 -- # rpc_cmd bdev_null_create null0 1024 512 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:47.507 null0 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@28 -- # rpc_cmd bdev_wait_for_examine 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@29 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 -a 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@30 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 null0 -g 16a3735b782d4338a53f5f5fc9d56ad1 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@31 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:47.507 [2024-07-15 18:07:41.069161] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@37 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -f ipv4 -s 4420 -n nqn.2016-06.io.spdk:cnode0 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:47.507 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:47.765 nvme0n1 00:21:47.765 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:47.765 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@41 -- # rpc_cmd bdev_get_bdevs -b nvme0n1 00:21:47.765 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:47.765 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:47.765 [ 00:21:47.765 { 00:21:47.765 "name": "nvme0n1", 00:21:47.765 "aliases": [ 00:21:47.765 "16a3735b-782d-4338-a53f-5f5fc9d56ad1" 00:21:47.765 ], 00:21:47.765 "product_name": "NVMe disk", 00:21:47.765 "block_size": 512, 00:21:47.765 "num_blocks": 2097152, 00:21:47.765 "uuid": "16a3735b-782d-4338-a53f-5f5fc9d56ad1", 00:21:47.765 "assigned_rate_limits": { 00:21:47.765 "rw_ios_per_sec": 0, 00:21:47.765 "rw_mbytes_per_sec": 0, 00:21:47.765 "r_mbytes_per_sec": 0, 00:21:47.765 "w_mbytes_per_sec": 0 00:21:47.765 }, 00:21:47.765 "claimed": false, 00:21:47.765 "zoned": false, 00:21:47.765 "supported_io_types": { 00:21:47.765 "read": true, 00:21:47.765 "write": true, 00:21:47.765 "unmap": false, 00:21:47.765 "flush": true, 00:21:47.765 "reset": true, 00:21:47.765 "nvme_admin": true, 00:21:47.765 "nvme_io": true, 00:21:47.765 "nvme_io_md": false, 00:21:47.765 "write_zeroes": true, 00:21:47.765 "zcopy": false, 00:21:47.765 "get_zone_info": false, 00:21:47.765 "zone_management": false, 00:21:47.765 "zone_append": false, 00:21:47.765 "compare": true, 00:21:47.765 "compare_and_write": true, 00:21:47.765 "abort": true, 00:21:47.765 "seek_hole": false, 00:21:47.765 "seek_data": false, 00:21:47.765 "copy": true, 00:21:47.765 "nvme_iov_md": false 00:21:47.765 }, 00:21:47.765 "memory_domains": [ 00:21:47.765 { 00:21:47.765 "dma_device_id": "system", 00:21:47.765 "dma_device_type": 1 00:21:47.765 } 00:21:47.765 ], 00:21:47.765 "driver_specific": { 00:21:47.765 "nvme": [ 00:21:47.765 { 00:21:47.765 "trid": { 00:21:47.765 "trtype": "TCP", 00:21:47.765 "adrfam": "IPv4", 00:21:47.765 "traddr": "10.0.0.2", 00:21:47.765 "trsvcid": "4420", 00:21:47.766 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:21:47.766 }, 00:21:47.766 "ctrlr_data": { 00:21:47.766 "cntlid": 1, 00:21:47.766 "vendor_id": "0x8086", 00:21:47.766 "model_number": "SPDK bdev Controller", 00:21:47.766 "serial_number": "00000000000000000000", 00:21:47.766 "firmware_revision": "24.09", 00:21:47.766 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:21:47.766 "oacs": { 00:21:47.766 "security": 0, 00:21:47.766 "format": 0, 00:21:47.766 "firmware": 0, 00:21:47.766 "ns_manage": 0 00:21:47.766 }, 00:21:47.766 "multi_ctrlr": true, 00:21:47.766 "ana_reporting": false 00:21:47.766 }, 00:21:47.766 "vs": { 00:21:47.766 "nvme_version": "1.3" 00:21:47.766 }, 00:21:47.766 "ns_data": { 00:21:47.766 "id": 1, 00:21:47.766 "can_share": true 00:21:47.766 } 00:21:47.766 } 00:21:47.766 ], 00:21:47.766 "mp_policy": "active_passive" 00:21:47.766 } 00:21:47.766 } 00:21:47.766 ] 00:21:47.766 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:47.766 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@44 -- # rpc_cmd bdev_nvme_reset_controller nvme0 00:21:47.766 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:47.766 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:47.766 [2024-07-15 18:07:41.317699] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:21:47.766 [2024-07-15 18:07:41.317752] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x12972d0 (9): Bad file descriptor 00:21:47.766 [2024-07-15 18:07:41.449302] bdev_nvme.c:2067:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:21:47.766 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:47.766 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@47 -- # rpc_cmd bdev_get_bdevs -b nvme0n1 00:21:47.766 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:47.766 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:47.766 [ 00:21:47.766 { 00:21:47.766 "name": "nvme0n1", 00:21:47.766 "aliases": [ 00:21:47.766 "16a3735b-782d-4338-a53f-5f5fc9d56ad1" 00:21:47.766 ], 00:21:47.766 "product_name": "NVMe disk", 00:21:47.766 "block_size": 512, 00:21:47.766 "num_blocks": 2097152, 00:21:47.766 "uuid": "16a3735b-782d-4338-a53f-5f5fc9d56ad1", 00:21:47.766 "assigned_rate_limits": { 00:21:47.766 "rw_ios_per_sec": 0, 00:21:47.766 "rw_mbytes_per_sec": 0, 00:21:47.766 "r_mbytes_per_sec": 0, 00:21:47.766 "w_mbytes_per_sec": 0 00:21:47.766 }, 00:21:47.766 "claimed": false, 00:21:47.766 "zoned": false, 00:21:47.766 "supported_io_types": { 00:21:47.766 "read": true, 00:21:47.766 "write": true, 00:21:47.766 "unmap": false, 00:21:47.766 "flush": true, 00:21:47.766 "reset": true, 00:21:47.766 "nvme_admin": true, 00:21:47.766 "nvme_io": true, 00:21:47.766 "nvme_io_md": false, 00:21:47.766 "write_zeroes": true, 00:21:47.766 "zcopy": false, 00:21:47.766 "get_zone_info": false, 00:21:47.766 "zone_management": false, 00:21:47.766 "zone_append": false, 00:21:47.766 "compare": true, 00:21:47.766 "compare_and_write": true, 00:21:47.766 "abort": true, 00:21:47.766 "seek_hole": false, 00:21:47.766 "seek_data": false, 00:21:47.766 "copy": true, 00:21:47.766 "nvme_iov_md": false 00:21:47.766 }, 00:21:47.766 "memory_domains": [ 00:21:47.766 { 00:21:47.766 "dma_device_id": "system", 00:21:47.766 "dma_device_type": 1 00:21:47.766 } 00:21:47.766 ], 00:21:47.766 "driver_specific": { 00:21:47.766 "nvme": [ 00:21:47.766 { 00:21:47.766 "trid": { 00:21:47.766 "trtype": "TCP", 00:21:47.766 "adrfam": "IPv4", 00:21:47.766 "traddr": "10.0.0.2", 00:21:47.766 "trsvcid": "4420", 00:21:47.766 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:21:47.766 }, 00:21:47.766 "ctrlr_data": { 00:21:47.766 "cntlid": 2, 00:21:47.766 "vendor_id": "0x8086", 00:21:47.766 "model_number": "SPDK bdev Controller", 00:21:47.766 "serial_number": "00000000000000000000", 00:21:47.766 "firmware_revision": "24.09", 00:21:47.766 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:21:47.766 "oacs": { 00:21:47.766 "security": 0, 00:21:47.766 "format": 0, 00:21:47.766 "firmware": 0, 00:21:47.766 "ns_manage": 0 00:21:47.766 }, 00:21:47.766 "multi_ctrlr": true, 00:21:47.766 "ana_reporting": false 00:21:47.766 }, 00:21:47.766 "vs": { 00:21:47.766 "nvme_version": "1.3" 00:21:47.766 }, 00:21:47.766 "ns_data": { 00:21:47.766 "id": 1, 00:21:47.766 "can_share": true 00:21:47.766 } 00:21:47.766 } 00:21:47.766 ], 00:21:47.766 "mp_policy": "active_passive" 00:21:47.766 } 00:21:47.766 } 00:21:47.766 ] 00:21:47.766 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:47.766 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@50 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:21:47.766 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:47.766 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:47.766 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:47.766 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@53 -- # mktemp 00:21:47.766 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@53 -- # key_path=/tmp/tmp.f0KpoLsAtu 00:21:47.766 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@54 -- # echo -n NVMeTLSkey-1:01:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: 00:21:47.766 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@55 -- # chmod 0600 /tmp/tmp.f0KpoLsAtu 00:21:47.766 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@56 -- # rpc_cmd nvmf_subsystem_allow_any_host nqn.2016-06.io.spdk:cnode0 --disable 00:21:47.766 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:47.766 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:48.024 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:48.024 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@57 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4421 --secure-channel 00:21:48.024 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:48.024 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:48.024 [2024-07-15 18:07:41.498256] tcp.c: 966:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:21:48.024 [2024-07-15 18:07:41.498354] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:21:48.024 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:48.024 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@59 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.f0KpoLsAtu 00:21:48.024 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:48.024 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:48.024 [2024-07-15 18:07:41.506272] tcp.c:3710:nvmf_tcp_subsystem_add_host: *WARNING*: nvmf_tcp_psk_path: deprecated feature PSK path to be removed in v24.09 00:21:48.024 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:48.024 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@65 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 10.0.0.2 -f ipv4 -s 4421 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host1 --psk /tmp/tmp.f0KpoLsAtu 00:21:48.024 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:48.024 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:48.024 [2024-07-15 18:07:41.514308] bdev_nvme_rpc.c: 517:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:21:48.024 [2024-07-15 18:07:41.514341] nvme_tcp.c:2589:nvme_tcp_generate_tls_credentials: *WARNING*: nvme_ctrlr_psk: deprecated feature spdk_nvme_ctrlr_opts.psk to be removed in v24.09 00:21:48.024 nvme0n1 00:21:48.024 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:48.024 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@69 -- # rpc_cmd bdev_get_bdevs -b nvme0n1 00:21:48.024 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:48.024 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:48.024 [ 00:21:48.024 { 00:21:48.024 "name": "nvme0n1", 00:21:48.024 "aliases": [ 00:21:48.024 "16a3735b-782d-4338-a53f-5f5fc9d56ad1" 00:21:48.024 ], 00:21:48.024 "product_name": "NVMe disk", 00:21:48.024 "block_size": 512, 00:21:48.024 "num_blocks": 2097152, 00:21:48.024 "uuid": "16a3735b-782d-4338-a53f-5f5fc9d56ad1", 00:21:48.024 "assigned_rate_limits": { 00:21:48.024 "rw_ios_per_sec": 0, 00:21:48.024 "rw_mbytes_per_sec": 0, 00:21:48.024 "r_mbytes_per_sec": 0, 00:21:48.024 "w_mbytes_per_sec": 0 00:21:48.024 }, 00:21:48.024 "claimed": false, 00:21:48.024 "zoned": false, 00:21:48.024 "supported_io_types": { 00:21:48.024 "read": true, 00:21:48.024 "write": true, 00:21:48.024 "unmap": false, 00:21:48.024 "flush": true, 00:21:48.024 "reset": true, 00:21:48.024 "nvme_admin": true, 00:21:48.024 "nvme_io": true, 00:21:48.024 "nvme_io_md": false, 00:21:48.024 "write_zeroes": true, 00:21:48.024 "zcopy": false, 00:21:48.024 "get_zone_info": false, 00:21:48.024 "zone_management": false, 00:21:48.024 "zone_append": false, 00:21:48.024 "compare": true, 00:21:48.024 "compare_and_write": true, 00:21:48.024 "abort": true, 00:21:48.025 "seek_hole": false, 00:21:48.025 "seek_data": false, 00:21:48.025 "copy": true, 00:21:48.025 "nvme_iov_md": false 00:21:48.025 }, 00:21:48.025 "memory_domains": [ 00:21:48.025 { 00:21:48.025 "dma_device_id": "system", 00:21:48.025 "dma_device_type": 1 00:21:48.025 } 00:21:48.025 ], 00:21:48.025 "driver_specific": { 00:21:48.025 "nvme": [ 00:21:48.025 { 00:21:48.025 "trid": { 00:21:48.025 "trtype": "TCP", 00:21:48.025 "adrfam": "IPv4", 00:21:48.025 "traddr": "10.0.0.2", 00:21:48.025 "trsvcid": "4421", 00:21:48.025 "subnqn": "nqn.2016-06.io.spdk:cnode0" 00:21:48.025 }, 00:21:48.025 "ctrlr_data": { 00:21:48.025 "cntlid": 3, 00:21:48.025 "vendor_id": "0x8086", 00:21:48.025 "model_number": "SPDK bdev Controller", 00:21:48.025 "serial_number": "00000000000000000000", 00:21:48.025 "firmware_revision": "24.09", 00:21:48.025 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:21:48.025 "oacs": { 00:21:48.025 "security": 0, 00:21:48.025 "format": 0, 00:21:48.025 "firmware": 0, 00:21:48.025 "ns_manage": 0 00:21:48.025 }, 00:21:48.025 "multi_ctrlr": true, 00:21:48.025 "ana_reporting": false 00:21:48.025 }, 00:21:48.025 "vs": { 00:21:48.025 "nvme_version": "1.3" 00:21:48.025 }, 00:21:48.025 "ns_data": { 00:21:48.025 "id": 1, 00:21:48.025 "can_share": true 00:21:48.025 } 00:21:48.025 } 00:21:48.025 ], 00:21:48.025 "mp_policy": "active_passive" 00:21:48.025 } 00:21:48.025 } 00:21:48.025 ] 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@72 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@75 -- # rm -f /tmp/tmp.f0KpoLsAtu 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@77 -- # trap - SIGINT SIGTERM EXIT 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- host/async_init.sh@78 -- # nvmftestfini 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@488 -- # nvmfcleanup 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@117 -- # sync 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@120 -- # set +e 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@121 -- # for i in {1..20} 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:21:48.025 rmmod nvme_tcp 00:21:48.025 rmmod nvme_fabrics 00:21:48.025 rmmod nvme_keyring 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@124 -- # set -e 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@125 -- # return 0 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@489 -- # '[' -n 664066 ']' 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@490 -- # killprocess 664066 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@948 -- # '[' -z 664066 ']' 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@952 -- # kill -0 664066 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@953 -- # uname 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 664066 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@966 -- # echo 'killing process with pid 664066' 00:21:48.025 killing process with pid 664066 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@967 -- # kill 664066 00:21:48.025 [2024-07-15 18:07:41.702822] app.c:1024:log_deprecation_hits: *WARNING*: nvme_ctrlr_psk: deprecation 'spdk_nvme_ctrlr_opts.psk' scheduled for removal in v24.09 hit 1 times 00:21:48.025 [2024-07-15 18:07:41.702845] app.c:1024:log_deprecation_hits: *WARNING*: nvmf_tcp_psk_path: deprecation 'PSK path' scheduled for removal in v24.09 hit 1 times 00:21:48.025 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@972 -- # wait 664066 00:21:48.282 18:07:41 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:21:48.282 18:07:41 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:21:48.282 18:07:41 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:21:48.282 18:07:41 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:21:48.282 18:07:41 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@278 -- # remove_spdk_ns 00:21:48.282 18:07:41 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:48.282 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:48.282 18:07:41 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:50.816 18:07:43 nvmf_tcp.nvmf_async_init -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:21:50.816 00:21:50.816 real 0m8.651s 00:21:50.816 user 0m3.132s 00:21:50.816 sys 0m3.887s 00:21:50.816 18:07:43 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@1124 -- # xtrace_disable 00:21:50.816 18:07:43 nvmf_tcp.nvmf_async_init -- common/autotest_common.sh@10 -- # set +x 00:21:50.816 ************************************ 00:21:50.816 END TEST nvmf_async_init 00:21:50.816 ************************************ 00:21:50.816 18:07:43 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:21:50.816 18:07:43 nvmf_tcp -- nvmf/nvmf.sh@94 -- # run_test dma /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/dma.sh --transport=tcp 00:21:50.816 18:07:43 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:21:50.816 18:07:43 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:21:50.816 18:07:43 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:21:50.816 ************************************ 00:21:50.816 START TEST dma 00:21:50.816 ************************************ 00:21:50.817 18:07:44 nvmf_tcp.dma -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/dma.sh --transport=tcp 00:21:50.817 * Looking for test storage... 00:21:50.817 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:21:50.817 18:07:44 nvmf_tcp.dma -- host/dma.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@7 -- # uname -s 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:21:50.817 18:07:44 nvmf_tcp.dma -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:21:50.817 18:07:44 nvmf_tcp.dma -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:21:50.817 18:07:44 nvmf_tcp.dma -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:21:50.817 18:07:44 nvmf_tcp.dma -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:50.817 18:07:44 nvmf_tcp.dma -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:50.817 18:07:44 nvmf_tcp.dma -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:50.817 18:07:44 nvmf_tcp.dma -- paths/export.sh@5 -- # export PATH 00:21:50.817 18:07:44 nvmf_tcp.dma -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@47 -- # : 0 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:21:50.817 18:07:44 nvmf_tcp.dma -- nvmf/common.sh@51 -- # have_pci_nics=0 00:21:50.817 18:07:44 nvmf_tcp.dma -- host/dma.sh@12 -- # '[' tcp '!=' rdma ']' 00:21:50.817 18:07:44 nvmf_tcp.dma -- host/dma.sh@13 -- # exit 0 00:21:50.817 00:21:50.817 real 0m0.115s 00:21:50.817 user 0m0.058s 00:21:50.817 sys 0m0.062s 00:21:50.817 18:07:44 nvmf_tcp.dma -- common/autotest_common.sh@1124 -- # xtrace_disable 00:21:50.817 18:07:44 nvmf_tcp.dma -- common/autotest_common.sh@10 -- # set +x 00:21:50.817 ************************************ 00:21:50.817 END TEST dma 00:21:50.817 ************************************ 00:21:50.817 18:07:44 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:21:50.817 18:07:44 nvmf_tcp -- nvmf/nvmf.sh@97 -- # run_test nvmf_identify /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify.sh --transport=tcp 00:21:50.817 18:07:44 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:21:50.817 18:07:44 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:21:50.817 18:07:44 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:21:50.817 ************************************ 00:21:50.817 START TEST nvmf_identify 00:21:50.817 ************************************ 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify.sh --transport=tcp 00:21:50.817 * Looking for test storage... 00:21:50.817 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- host/identify.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@7 -- # uname -s 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- paths/export.sh@5 -- # export PATH 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@47 -- # : 0 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@51 -- # have_pci_nics=0 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- host/identify.sh@11 -- # MALLOC_BDEV_SIZE=64 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- host/identify.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- host/identify.sh@14 -- # nvmftestinit 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:21:50.817 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:21:50.818 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@448 -- # prepare_net_devs 00:21:50.818 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@410 -- # local -g is_hw=no 00:21:50.818 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@412 -- # remove_spdk_ns 00:21:50.818 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:50.818 18:07:44 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:50.818 18:07:44 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:50.818 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:21:50.818 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:21:50.818 18:07:44 nvmf_tcp.nvmf_identify -- nvmf/common.sh@285 -- # xtrace_disable 00:21:50.818 18:07:44 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@291 -- # pci_devs=() 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@291 -- # local -a pci_devs 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@292 -- # pci_net_devs=() 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@293 -- # pci_drivers=() 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@293 -- # local -A pci_drivers 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@295 -- # net_devs=() 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@295 -- # local -ga net_devs 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@296 -- # e810=() 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@296 -- # local -ga e810 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@297 -- # x722=() 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@297 -- # local -ga x722 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@298 -- # mlx=() 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@298 -- # local -ga mlx 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:21:56.089 Found 0000:86:00.0 (0x8086 - 0x159b) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:21:56.089 Found 0000:86:00.1 (0x8086 - 0x159b) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@390 -- # [[ up == up ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:21:56.089 Found net devices under 0000:86:00.0: cvl_0_0 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@390 -- # [[ up == up ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:21:56.089 Found net devices under 0000:86:00.1: cvl_0_1 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@414 -- # is_hw=yes 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:21:56.089 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:21:56.090 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:21:56.090 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.157 ms 00:21:56.090 00:21:56.090 --- 10.0.0.2 ping statistics --- 00:21:56.090 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:56.090 rtt min/avg/max/mdev = 0.157/0.157/0.157/0.000 ms 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:21:56.090 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:21:56.090 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.231 ms 00:21:56.090 00:21:56.090 --- 10.0.0.1 ping statistics --- 00:21:56.090 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:21:56.090 rtt min/avg/max/mdev = 0.231/0.231/0.231/0.000 ms 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@422 -- # return 0 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- host/identify.sh@16 -- # timing_enter start_nvmf_tgt 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@722 -- # xtrace_disable 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- host/identify.sh@19 -- # nvmfpid=667870 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- host/identify.sh@21 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- host/identify.sh@18 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- host/identify.sh@23 -- # waitforlisten 667870 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@829 -- # '[' -z 667870 ']' 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@834 -- # local max_retries=100 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:21:56.090 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@838 -- # xtrace_disable 00:21:56.090 18:07:49 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:21:56.090 [2024-07-15 18:07:49.767212] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:21:56.090 [2024-07-15 18:07:49.767263] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:21:56.090 EAL: No free 2048 kB hugepages reported on node 1 00:21:56.349 [2024-07-15 18:07:49.825413] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:21:56.349 [2024-07-15 18:07:49.905331] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:21:56.349 [2024-07-15 18:07:49.905369] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:21:56.349 [2024-07-15 18:07:49.905377] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:21:56.349 [2024-07-15 18:07:49.905383] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:21:56.349 [2024-07-15 18:07:49.905388] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:21:56.349 [2024-07-15 18:07:49.905430] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:21:56.349 [2024-07-15 18:07:49.905525] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:21:56.349 [2024-07-15 18:07:49.905613] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:21:56.349 [2024-07-15 18:07:49.905615] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:21:56.917 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:21:56.917 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@862 -- # return 0 00:21:56.917 18:07:50 nvmf_tcp.nvmf_identify -- host/identify.sh@24 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:21:56.917 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:56.917 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:21:56.917 [2024-07-15 18:07:50.586215] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:21:56.917 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:56.917 18:07:50 nvmf_tcp.nvmf_identify -- host/identify.sh@25 -- # timing_exit start_nvmf_tgt 00:21:56.917 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@728 -- # xtrace_disable 00:21:56.917 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:21:56.917 18:07:50 nvmf_tcp.nvmf_identify -- host/identify.sh@27 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:21:56.917 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:56.917 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:21:57.179 Malloc0 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- host/identify.sh@28 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- host/identify.sh@31 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 --nguid ABCDEF0123456789ABCDEF0123456789 --eui64 ABCDEF0123456789 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- host/identify.sh@34 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:21:57.179 [2024-07-15 18:07:50.666220] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- host/identify.sh@35 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- host/identify.sh@37 -- # rpc_cmd nvmf_get_subsystems 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:21:57.179 [ 00:21:57.179 { 00:21:57.179 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:21:57.179 "subtype": "Discovery", 00:21:57.179 "listen_addresses": [ 00:21:57.179 { 00:21:57.179 "trtype": "TCP", 00:21:57.179 "adrfam": "IPv4", 00:21:57.179 "traddr": "10.0.0.2", 00:21:57.179 "trsvcid": "4420" 00:21:57.179 } 00:21:57.179 ], 00:21:57.179 "allow_any_host": true, 00:21:57.179 "hosts": [] 00:21:57.179 }, 00:21:57.179 { 00:21:57.179 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:21:57.179 "subtype": "NVMe", 00:21:57.179 "listen_addresses": [ 00:21:57.179 { 00:21:57.179 "trtype": "TCP", 00:21:57.179 "adrfam": "IPv4", 00:21:57.179 "traddr": "10.0.0.2", 00:21:57.179 "trsvcid": "4420" 00:21:57.179 } 00:21:57.179 ], 00:21:57.179 "allow_any_host": true, 00:21:57.179 "hosts": [], 00:21:57.179 "serial_number": "SPDK00000000000001", 00:21:57.179 "model_number": "SPDK bdev Controller", 00:21:57.179 "max_namespaces": 32, 00:21:57.179 "min_cntlid": 1, 00:21:57.179 "max_cntlid": 65519, 00:21:57.179 "namespaces": [ 00:21:57.179 { 00:21:57.179 "nsid": 1, 00:21:57.179 "bdev_name": "Malloc0", 00:21:57.179 "name": "Malloc0", 00:21:57.179 "nguid": "ABCDEF0123456789ABCDEF0123456789", 00:21:57.179 "eui64": "ABCDEF0123456789", 00:21:57.179 "uuid": "98dd49eb-a0db-43a6-9107-626bc56df2fd" 00:21:57.179 } 00:21:57.179 ] 00:21:57.179 } 00:21:57.179 ] 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:57.179 18:07:50 nvmf_tcp.nvmf_identify -- host/identify.sh@39 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2014-08.org.nvmexpress.discovery' -L all 00:21:57.179 [2024-07-15 18:07:50.717609] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:21:57.179 [2024-07-15 18:07:50.717643] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid668115 ] 00:21:57.179 EAL: No free 2048 kB hugepages reported on node 1 00:21:57.179 [2024-07-15 18:07:50.746728] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to connect adminq (no timeout) 00:21:57.179 [2024-07-15 18:07:50.746779] nvme_tcp.c:2338:nvme_tcp_qpair_connect_sock: *DEBUG*: adrfam 1 ai_family 2 00:21:57.179 [2024-07-15 18:07:50.746784] nvme_tcp.c:2342:nvme_tcp_qpair_connect_sock: *DEBUG*: trsvcid is 4420 00:21:57.179 [2024-07-15 18:07:50.746796] nvme_tcp.c:2360:nvme_tcp_qpair_connect_sock: *DEBUG*: sock_impl_name is (null) 00:21:57.179 [2024-07-15 18:07:50.746803] sock.c: 337:spdk_sock_connect_ext: *DEBUG*: Creating a client socket using impl posix 00:21:57.179 [2024-07-15 18:07:50.747104] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to wait for connect adminq (no timeout) 00:21:57.179 [2024-07-15 18:07:50.747134] nvme_tcp.c:1555:nvme_tcp_send_icreq_complete: *DEBUG*: Complete the icreq send for tqpair=0x1011ec0 0 00:21:57.180 [2024-07-15 18:07:50.761234] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 1 00:21:57.180 [2024-07-15 18:07:50.761254] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =1 00:21:57.180 [2024-07-15 18:07:50.761259] nvme_tcp.c:1601:nvme_tcp_icresp_handle: *DEBUG*: host_hdgst_enable: 0 00:21:57.180 [2024-07-15 18:07:50.761262] nvme_tcp.c:1602:nvme_tcp_icresp_handle: *DEBUG*: host_ddgst_enable: 0 00:21:57.180 [2024-07-15 18:07:50.761299] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.761304] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.761308] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1011ec0) 00:21:57.180 [2024-07-15 18:07:50.761321] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:0 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x400 00:21:57.180 [2024-07-15 18:07:50.761337] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1094e40, cid 0, qid 0 00:21:57.180 [2024-07-15 18:07:50.769235] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.180 [2024-07-15 18:07:50.769243] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.180 [2024-07-15 18:07:50.769246] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.769250] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1094e40) on tqpair=0x1011ec0 00:21:57.180 [2024-07-15 18:07:50.769260] nvme_fabric.c: 622:_nvme_fabric_qpair_connect_poll: *DEBUG*: CNTLID 0x0001 00:21:57.180 [2024-07-15 18:07:50.769266] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to read vs (no timeout) 00:21:57.180 [2024-07-15 18:07:50.769271] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to read vs wait for vs (no timeout) 00:21:57.180 [2024-07-15 18:07:50.769284] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.769288] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.769291] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1011ec0) 00:21:57.180 [2024-07-15 18:07:50.769298] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.180 [2024-07-15 18:07:50.769311] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1094e40, cid 0, qid 0 00:21:57.180 [2024-07-15 18:07:50.769479] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.180 [2024-07-15 18:07:50.769486] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.180 [2024-07-15 18:07:50.769489] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.769493] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1094e40) on tqpair=0x1011ec0 00:21:57.180 [2024-07-15 18:07:50.769497] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to read cap (no timeout) 00:21:57.180 [2024-07-15 18:07:50.769504] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to read cap wait for cap (no timeout) 00:21:57.180 [2024-07-15 18:07:50.769510] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.769514] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.769517] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1011ec0) 00:21:57.180 [2024-07-15 18:07:50.769523] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.180 [2024-07-15 18:07:50.769534] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1094e40, cid 0, qid 0 00:21:57.180 [2024-07-15 18:07:50.769606] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.180 [2024-07-15 18:07:50.769611] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.180 [2024-07-15 18:07:50.769614] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.769618] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1094e40) on tqpair=0x1011ec0 00:21:57.180 [2024-07-15 18:07:50.769623] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to check en (no timeout) 00:21:57.180 [2024-07-15 18:07:50.769629] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to check en wait for cc (timeout 15000 ms) 00:21:57.180 [2024-07-15 18:07:50.769635] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.769639] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.769642] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1011ec0) 00:21:57.180 [2024-07-15 18:07:50.769648] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.180 [2024-07-15 18:07:50.769657] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1094e40, cid 0, qid 0 00:21:57.180 [2024-07-15 18:07:50.769728] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.180 [2024-07-15 18:07:50.769734] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.180 [2024-07-15 18:07:50.769737] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.769741] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1094e40) on tqpair=0x1011ec0 00:21:57.180 [2024-07-15 18:07:50.769745] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:21:57.180 [2024-07-15 18:07:50.769753] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.769757] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.769760] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1011ec0) 00:21:57.180 [2024-07-15 18:07:50.769766] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.180 [2024-07-15 18:07:50.769775] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1094e40, cid 0, qid 0 00:21:57.180 [2024-07-15 18:07:50.769848] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.180 [2024-07-15 18:07:50.769854] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.180 [2024-07-15 18:07:50.769859] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.769862] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1094e40) on tqpair=0x1011ec0 00:21:57.180 [2024-07-15 18:07:50.769866] nvme_ctrlr.c:3869:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] CC.EN = 0 && CSTS.RDY = 0 00:21:57.180 [2024-07-15 18:07:50.769870] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to controller is disabled (timeout 15000 ms) 00:21:57.180 [2024-07-15 18:07:50.769877] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:21:57.180 [2024-07-15 18:07:50.769982] nvme_ctrlr.c:4062:nvme_ctrlr_process_init: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] Setting CC.EN = 1 00:21:57.180 [2024-07-15 18:07:50.769986] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:21:57.180 [2024-07-15 18:07:50.769994] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.769997] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.770000] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1011ec0) 00:21:57.180 [2024-07-15 18:07:50.770006] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.180 [2024-07-15 18:07:50.770016] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1094e40, cid 0, qid 0 00:21:57.180 [2024-07-15 18:07:50.770100] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.180 [2024-07-15 18:07:50.770106] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.180 [2024-07-15 18:07:50.770109] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.770113] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1094e40) on tqpair=0x1011ec0 00:21:57.180 [2024-07-15 18:07:50.770117] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:21:57.180 [2024-07-15 18:07:50.770124] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.770128] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.770131] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1011ec0) 00:21:57.180 [2024-07-15 18:07:50.770137] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.180 [2024-07-15 18:07:50.770146] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1094e40, cid 0, qid 0 00:21:57.180 [2024-07-15 18:07:50.770221] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.180 [2024-07-15 18:07:50.770231] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.180 [2024-07-15 18:07:50.770235] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.770238] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1094e40) on tqpair=0x1011ec0 00:21:57.180 [2024-07-15 18:07:50.770242] nvme_ctrlr.c:3904:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:21:57.180 [2024-07-15 18:07:50.770246] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to reset admin queue (timeout 30000 ms) 00:21:57.180 [2024-07-15 18:07:50.770253] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to identify controller (no timeout) 00:21:57.180 [2024-07-15 18:07:50.770261] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to wait for identify controller (timeout 30000 ms) 00:21:57.180 [2024-07-15 18:07:50.770270] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.770273] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1011ec0) 00:21:57.180 [2024-07-15 18:07:50.770282] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:0 nsid:0 cdw10:00000001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.180 [2024-07-15 18:07:50.770292] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1094e40, cid 0, qid 0 00:21:57.180 [2024-07-15 18:07:50.770400] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:57.180 [2024-07-15 18:07:50.770406] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:57.180 [2024-07-15 18:07:50.770409] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.770413] nvme_tcp.c:1720:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1011ec0): datao=0, datal=4096, cccid=0 00:21:57.180 [2024-07-15 18:07:50.770417] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1094e40) on tqpair(0x1011ec0): expected_datao=0, payload_size=4096 00:21:57.180 [2024-07-15 18:07:50.770420] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.770445] nvme_tcp.c:1521:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.770449] nvme_tcp.c:1312:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.814232] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.180 [2024-07-15 18:07:50.814244] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.180 [2024-07-15 18:07:50.814248] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.180 [2024-07-15 18:07:50.814251] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1094e40) on tqpair=0x1011ec0 00:21:57.180 [2024-07-15 18:07:50.814259] nvme_ctrlr.c:2053:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] transport max_xfer_size 4294967295 00:21:57.180 [2024-07-15 18:07:50.814267] nvme_ctrlr.c:2057:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] MDTS max_xfer_size 131072 00:21:57.180 [2024-07-15 18:07:50.814271] nvme_ctrlr.c:2060:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] CNTLID 0x0001 00:21:57.180 [2024-07-15 18:07:50.814276] nvme_ctrlr.c:2084:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] transport max_sges 16 00:21:57.180 [2024-07-15 18:07:50.814280] nvme_ctrlr.c:2099:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] fuses compare and write: 1 00:21:57.180 [2024-07-15 18:07:50.814284] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to configure AER (timeout 30000 ms) 00:21:57.181 [2024-07-15 18:07:50.814293] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to wait for configure aer (timeout 30000 ms) 00:21:57.181 [2024-07-15 18:07:50.814300] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814304] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814307] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1011ec0) 00:21:57.181 [2024-07-15 18:07:50.814314] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:0 cdw10:0000000b SGL DATA BLOCK OFFSET 0x0 len:0x0 00:21:57.181 [2024-07-15 18:07:50.814327] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1094e40, cid 0, qid 0 00:21:57.181 [2024-07-15 18:07:50.814487] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.181 [2024-07-15 18:07:50.814493] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.181 [2024-07-15 18:07:50.814496] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814500] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1094e40) on tqpair=0x1011ec0 00:21:57.181 [2024-07-15 18:07:50.814506] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814510] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814513] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x1011ec0) 00:21:57.181 [2024-07-15 18:07:50.814522] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:57.181 [2024-07-15 18:07:50.814527] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814531] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814534] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=1 on tqpair(0x1011ec0) 00:21:57.181 [2024-07-15 18:07:50.814539] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:57.181 [2024-07-15 18:07:50.814544] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814548] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814551] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=2 on tqpair(0x1011ec0) 00:21:57.181 [2024-07-15 18:07:50.814556] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:57.181 [2024-07-15 18:07:50.814561] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814564] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814567] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1011ec0) 00:21:57.181 [2024-07-15 18:07:50.814572] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:57.181 [2024-07-15 18:07:50.814577] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to set keep alive timeout (timeout 30000 ms) 00:21:57.181 [2024-07-15 18:07:50.814587] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:21:57.181 [2024-07-15 18:07:50.814593] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814596] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1011ec0) 00:21:57.181 [2024-07-15 18:07:50.814603] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:4 cdw10:0000000f SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.181 [2024-07-15 18:07:50.814614] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1094e40, cid 0, qid 0 00:21:57.181 [2024-07-15 18:07:50.814619] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1094fc0, cid 1, qid 0 00:21:57.181 [2024-07-15 18:07:50.814623] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1095140, cid 2, qid 0 00:21:57.181 [2024-07-15 18:07:50.814627] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x10952c0, cid 3, qid 0 00:21:57.181 [2024-07-15 18:07:50.814631] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1095440, cid 4, qid 0 00:21:57.181 [2024-07-15 18:07:50.814746] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.181 [2024-07-15 18:07:50.814752] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.181 [2024-07-15 18:07:50.814755] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814759] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1095440) on tqpair=0x1011ec0 00:21:57.181 [2024-07-15 18:07:50.814763] nvme_ctrlr.c:3022:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] Sending keep alive every 5000000 us 00:21:57.181 [2024-07-15 18:07:50.814768] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] setting state to ready (no timeout) 00:21:57.181 [2024-07-15 18:07:50.814777] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814781] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1011ec0) 00:21:57.181 [2024-07-15 18:07:50.814787] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:0 cdw10:00000001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.181 [2024-07-15 18:07:50.814798] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1095440, cid 4, qid 0 00:21:57.181 [2024-07-15 18:07:50.814881] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:57.181 [2024-07-15 18:07:50.814887] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:57.181 [2024-07-15 18:07:50.814890] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814894] nvme_tcp.c:1720:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1011ec0): datao=0, datal=4096, cccid=4 00:21:57.181 [2024-07-15 18:07:50.814898] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1095440) on tqpair(0x1011ec0): expected_datao=0, payload_size=4096 00:21:57.181 [2024-07-15 18:07:50.814901] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814908] nvme_tcp.c:1521:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814911] nvme_tcp.c:1312:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814934] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.181 [2024-07-15 18:07:50.814940] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.181 [2024-07-15 18:07:50.814943] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814946] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1095440) on tqpair=0x1011ec0 00:21:57.181 [2024-07-15 18:07:50.814957] nvme_ctrlr.c:4160:nvme_ctrlr_process_init: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] Ctrlr already in ready state 00:21:57.181 [2024-07-15 18:07:50.814979] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814983] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1011ec0) 00:21:57.181 [2024-07-15 18:07:50.814989] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:0 cdw10:00ff0070 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.181 [2024-07-15 18:07:50.814995] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.814999] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.815002] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x1011ec0) 00:21:57.181 [2024-07-15 18:07:50.815007] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:5 nsid:0 cdw10:00000000 cdw11:00000000 00:21:57.181 [2024-07-15 18:07:50.815020] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1095440, cid 4, qid 0 00:21:57.181 [2024-07-15 18:07:50.815025] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x10955c0, cid 5, qid 0 00:21:57.181 [2024-07-15 18:07:50.815134] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:57.181 [2024-07-15 18:07:50.815140] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:57.181 [2024-07-15 18:07:50.815143] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.815147] nvme_tcp.c:1720:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1011ec0): datao=0, datal=1024, cccid=4 00:21:57.181 [2024-07-15 18:07:50.815151] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1095440) on tqpair(0x1011ec0): expected_datao=0, payload_size=1024 00:21:57.181 [2024-07-15 18:07:50.815154] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.815160] nvme_tcp.c:1521:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.815163] nvme_tcp.c:1312:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.815168] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.181 [2024-07-15 18:07:50.815173] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.181 [2024-07-15 18:07:50.815176] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.815180] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x10955c0) on tqpair=0x1011ec0 00:21:57.181 [2024-07-15 18:07:50.858234] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.181 [2024-07-15 18:07:50.858245] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.181 [2024-07-15 18:07:50.858251] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.858256] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1095440) on tqpair=0x1011ec0 00:21:57.181 [2024-07-15 18:07:50.858272] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.858276] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1011ec0) 00:21:57.181 [2024-07-15 18:07:50.858283] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:0 cdw10:02ff0070 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.181 [2024-07-15 18:07:50.858299] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1095440, cid 4, qid 0 00:21:57.181 [2024-07-15 18:07:50.858467] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:57.181 [2024-07-15 18:07:50.858473] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:57.181 [2024-07-15 18:07:50.858476] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.858480] nvme_tcp.c:1720:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1011ec0): datao=0, datal=3072, cccid=4 00:21:57.181 [2024-07-15 18:07:50.858484] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1095440) on tqpair(0x1011ec0): expected_datao=0, payload_size=3072 00:21:57.181 [2024-07-15 18:07:50.858487] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.858493] nvme_tcp.c:1521:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.858497] nvme_tcp.c:1312:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.858537] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.181 [2024-07-15 18:07:50.858542] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.181 [2024-07-15 18:07:50.858546] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.858549] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1095440) on tqpair=0x1011ec0 00:21:57.181 [2024-07-15 18:07:50.858557] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.858561] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x1011ec0) 00:21:57.181 [2024-07-15 18:07:50.858567] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:0 cdw10:00010070 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.181 [2024-07-15 18:07:50.858579] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x1095440, cid 4, qid 0 00:21:57.181 [2024-07-15 18:07:50.858665] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:57.181 [2024-07-15 18:07:50.858671] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:57.181 [2024-07-15 18:07:50.858674] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:57.181 [2024-07-15 18:07:50.858678] nvme_tcp.c:1720:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x1011ec0): datao=0, datal=8, cccid=4 00:21:57.181 [2024-07-15 18:07:50.858682] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x1095440) on tqpair(0x1011ec0): expected_datao=0, payload_size=8 00:21:57.182 [2024-07-15 18:07:50.858685] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.182 [2024-07-15 18:07:50.858691] nvme_tcp.c:1521:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:57.182 [2024-07-15 18:07:50.858694] nvme_tcp.c:1312:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:57.182 [2024-07-15 18:07:50.900370] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.182 [2024-07-15 18:07:50.900380] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.182 [2024-07-15 18:07:50.900383] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.182 [2024-07-15 18:07:50.900387] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1095440) on tqpair=0x1011ec0 00:21:57.182 ===================================================== 00:21:57.182 NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2014-08.org.nvmexpress.discovery 00:21:57.182 ===================================================== 00:21:57.182 Controller Capabilities/Features 00:21:57.182 ================================ 00:21:57.182 Vendor ID: 0000 00:21:57.182 Subsystem Vendor ID: 0000 00:21:57.182 Serial Number: .................... 00:21:57.182 Model Number: ........................................ 00:21:57.182 Firmware Version: 24.09 00:21:57.182 Recommended Arb Burst: 0 00:21:57.182 IEEE OUI Identifier: 00 00 00 00:21:57.182 Multi-path I/O 00:21:57.182 May have multiple subsystem ports: No 00:21:57.182 May have multiple controllers: No 00:21:57.182 Associated with SR-IOV VF: No 00:21:57.182 Max Data Transfer Size: 131072 00:21:57.182 Max Number of Namespaces: 0 00:21:57.182 Max Number of I/O Queues: 1024 00:21:57.182 NVMe Specification Version (VS): 1.3 00:21:57.182 NVMe Specification Version (Identify): 1.3 00:21:57.182 Maximum Queue Entries: 128 00:21:57.182 Contiguous Queues Required: Yes 00:21:57.182 Arbitration Mechanisms Supported 00:21:57.182 Weighted Round Robin: Not Supported 00:21:57.182 Vendor Specific: Not Supported 00:21:57.182 Reset Timeout: 15000 ms 00:21:57.182 Doorbell Stride: 4 bytes 00:21:57.182 NVM Subsystem Reset: Not Supported 00:21:57.182 Command Sets Supported 00:21:57.182 NVM Command Set: Supported 00:21:57.182 Boot Partition: Not Supported 00:21:57.182 Memory Page Size Minimum: 4096 bytes 00:21:57.182 Memory Page Size Maximum: 4096 bytes 00:21:57.182 Persistent Memory Region: Not Supported 00:21:57.182 Optional Asynchronous Events Supported 00:21:57.182 Namespace Attribute Notices: Not Supported 00:21:57.182 Firmware Activation Notices: Not Supported 00:21:57.182 ANA Change Notices: Not Supported 00:21:57.182 PLE Aggregate Log Change Notices: Not Supported 00:21:57.182 LBA Status Info Alert Notices: Not Supported 00:21:57.182 EGE Aggregate Log Change Notices: Not Supported 00:21:57.182 Normal NVM Subsystem Shutdown event: Not Supported 00:21:57.182 Zone Descriptor Change Notices: Not Supported 00:21:57.182 Discovery Log Change Notices: Supported 00:21:57.182 Controller Attributes 00:21:57.182 128-bit Host Identifier: Not Supported 00:21:57.182 Non-Operational Permissive Mode: Not Supported 00:21:57.182 NVM Sets: Not Supported 00:21:57.182 Read Recovery Levels: Not Supported 00:21:57.182 Endurance Groups: Not Supported 00:21:57.182 Predictable Latency Mode: Not Supported 00:21:57.182 Traffic Based Keep ALive: Not Supported 00:21:57.182 Namespace Granularity: Not Supported 00:21:57.182 SQ Associations: Not Supported 00:21:57.182 UUID List: Not Supported 00:21:57.182 Multi-Domain Subsystem: Not Supported 00:21:57.182 Fixed Capacity Management: Not Supported 00:21:57.182 Variable Capacity Management: Not Supported 00:21:57.182 Delete Endurance Group: Not Supported 00:21:57.182 Delete NVM Set: Not Supported 00:21:57.182 Extended LBA Formats Supported: Not Supported 00:21:57.182 Flexible Data Placement Supported: Not Supported 00:21:57.182 00:21:57.182 Controller Memory Buffer Support 00:21:57.182 ================================ 00:21:57.182 Supported: No 00:21:57.182 00:21:57.182 Persistent Memory Region Support 00:21:57.182 ================================ 00:21:57.182 Supported: No 00:21:57.182 00:21:57.182 Admin Command Set Attributes 00:21:57.182 ============================ 00:21:57.182 Security Send/Receive: Not Supported 00:21:57.182 Format NVM: Not Supported 00:21:57.182 Firmware Activate/Download: Not Supported 00:21:57.182 Namespace Management: Not Supported 00:21:57.182 Device Self-Test: Not Supported 00:21:57.182 Directives: Not Supported 00:21:57.182 NVMe-MI: Not Supported 00:21:57.182 Virtualization Management: Not Supported 00:21:57.182 Doorbell Buffer Config: Not Supported 00:21:57.182 Get LBA Status Capability: Not Supported 00:21:57.182 Command & Feature Lockdown Capability: Not Supported 00:21:57.182 Abort Command Limit: 1 00:21:57.182 Async Event Request Limit: 4 00:21:57.182 Number of Firmware Slots: N/A 00:21:57.182 Firmware Slot 1 Read-Only: N/A 00:21:57.182 Firmware Activation Without Reset: N/A 00:21:57.182 Multiple Update Detection Support: N/A 00:21:57.182 Firmware Update Granularity: No Information Provided 00:21:57.182 Per-Namespace SMART Log: No 00:21:57.182 Asymmetric Namespace Access Log Page: Not Supported 00:21:57.182 Subsystem NQN: nqn.2014-08.org.nvmexpress.discovery 00:21:57.182 Command Effects Log Page: Not Supported 00:21:57.182 Get Log Page Extended Data: Supported 00:21:57.182 Telemetry Log Pages: Not Supported 00:21:57.182 Persistent Event Log Pages: Not Supported 00:21:57.182 Supported Log Pages Log Page: May Support 00:21:57.182 Commands Supported & Effects Log Page: Not Supported 00:21:57.182 Feature Identifiers & Effects Log Page:May Support 00:21:57.182 NVMe-MI Commands & Effects Log Page: May Support 00:21:57.182 Data Area 4 for Telemetry Log: Not Supported 00:21:57.182 Error Log Page Entries Supported: 128 00:21:57.182 Keep Alive: Not Supported 00:21:57.182 00:21:57.182 NVM Command Set Attributes 00:21:57.182 ========================== 00:21:57.182 Submission Queue Entry Size 00:21:57.182 Max: 1 00:21:57.182 Min: 1 00:21:57.182 Completion Queue Entry Size 00:21:57.182 Max: 1 00:21:57.182 Min: 1 00:21:57.182 Number of Namespaces: 0 00:21:57.182 Compare Command: Not Supported 00:21:57.182 Write Uncorrectable Command: Not Supported 00:21:57.182 Dataset Management Command: Not Supported 00:21:57.182 Write Zeroes Command: Not Supported 00:21:57.182 Set Features Save Field: Not Supported 00:21:57.182 Reservations: Not Supported 00:21:57.182 Timestamp: Not Supported 00:21:57.182 Copy: Not Supported 00:21:57.182 Volatile Write Cache: Not Present 00:21:57.182 Atomic Write Unit (Normal): 1 00:21:57.182 Atomic Write Unit (PFail): 1 00:21:57.182 Atomic Compare & Write Unit: 1 00:21:57.182 Fused Compare & Write: Supported 00:21:57.182 Scatter-Gather List 00:21:57.182 SGL Command Set: Supported 00:21:57.182 SGL Keyed: Supported 00:21:57.182 SGL Bit Bucket Descriptor: Not Supported 00:21:57.182 SGL Metadata Pointer: Not Supported 00:21:57.182 Oversized SGL: Not Supported 00:21:57.182 SGL Metadata Address: Not Supported 00:21:57.182 SGL Offset: Supported 00:21:57.182 Transport SGL Data Block: Not Supported 00:21:57.182 Replay Protected Memory Block: Not Supported 00:21:57.182 00:21:57.182 Firmware Slot Information 00:21:57.182 ========================= 00:21:57.182 Active slot: 0 00:21:57.182 00:21:57.182 00:21:57.182 Error Log 00:21:57.182 ========= 00:21:57.182 00:21:57.182 Active Namespaces 00:21:57.182 ================= 00:21:57.182 Discovery Log Page 00:21:57.182 ================== 00:21:57.182 Generation Counter: 2 00:21:57.182 Number of Records: 2 00:21:57.182 Record Format: 0 00:21:57.182 00:21:57.182 Discovery Log Entry 0 00:21:57.182 ---------------------- 00:21:57.182 Transport Type: 3 (TCP) 00:21:57.182 Address Family: 1 (IPv4) 00:21:57.182 Subsystem Type: 3 (Current Discovery Subsystem) 00:21:57.182 Entry Flags: 00:21:57.182 Duplicate Returned Information: 1 00:21:57.182 Explicit Persistent Connection Support for Discovery: 1 00:21:57.182 Transport Requirements: 00:21:57.182 Secure Channel: Not Required 00:21:57.182 Port ID: 0 (0x0000) 00:21:57.182 Controller ID: 65535 (0xffff) 00:21:57.182 Admin Max SQ Size: 128 00:21:57.182 Transport Service Identifier: 4420 00:21:57.182 NVM Subsystem Qualified Name: nqn.2014-08.org.nvmexpress.discovery 00:21:57.182 Transport Address: 10.0.0.2 00:21:57.182 Discovery Log Entry 1 00:21:57.182 ---------------------- 00:21:57.182 Transport Type: 3 (TCP) 00:21:57.182 Address Family: 1 (IPv4) 00:21:57.182 Subsystem Type: 2 (NVM Subsystem) 00:21:57.182 Entry Flags: 00:21:57.182 Duplicate Returned Information: 0 00:21:57.182 Explicit Persistent Connection Support for Discovery: 0 00:21:57.182 Transport Requirements: 00:21:57.182 Secure Channel: Not Required 00:21:57.182 Port ID: 0 (0x0000) 00:21:57.182 Controller ID: 65535 (0xffff) 00:21:57.182 Admin Max SQ Size: 128 00:21:57.182 Transport Service Identifier: 4420 00:21:57.182 NVM Subsystem Qualified Name: nqn.2016-06.io.spdk:cnode1 00:21:57.182 Transport Address: 10.0.0.2 [2024-07-15 18:07:50.900464] nvme_ctrlr.c:4357:nvme_ctrlr_destruct_async: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] Prepare to destruct SSD 00:21:57.182 [2024-07-15 18:07:50.900474] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1094e40) on tqpair=0x1011ec0 00:21:57.183 [2024-07-15 18:07:50.900482] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:57.183 [2024-07-15 18:07:50.900487] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1094fc0) on tqpair=0x1011ec0 00:21:57.183 [2024-07-15 18:07:50.900491] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:57.183 [2024-07-15 18:07:50.900495] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x1095140) on tqpair=0x1011ec0 00:21:57.183 [2024-07-15 18:07:50.900499] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:57.183 [2024-07-15 18:07:50.900503] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x10952c0) on tqpair=0x1011ec0 00:21:57.183 [2024-07-15 18:07:50.900507] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:57.183 [2024-07-15 18:07:50.900517] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.900521] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.900524] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1011ec0) 00:21:57.183 [2024-07-15 18:07:50.900530] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.183 [2024-07-15 18:07:50.900543] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x10952c0, cid 3, qid 0 00:21:57.183 [2024-07-15 18:07:50.900614] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.183 [2024-07-15 18:07:50.900620] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.183 [2024-07-15 18:07:50.900623] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.900627] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x10952c0) on tqpair=0x1011ec0 00:21:57.183 [2024-07-15 18:07:50.900633] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.900636] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.900639] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1011ec0) 00:21:57.183 [2024-07-15 18:07:50.900645] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.183 [2024-07-15 18:07:50.900658] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x10952c0, cid 3, qid 0 00:21:57.183 [2024-07-15 18:07:50.900743] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.183 [2024-07-15 18:07:50.900749] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.183 [2024-07-15 18:07:50.900752] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.900755] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x10952c0) on tqpair=0x1011ec0 00:21:57.183 [2024-07-15 18:07:50.900759] nvme_ctrlr.c:1147:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] RTD3E = 0 us 00:21:57.183 [2024-07-15 18:07:50.900763] nvme_ctrlr.c:1150:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] shutdown timeout = 10000 ms 00:21:57.183 [2024-07-15 18:07:50.900771] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.900775] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.900778] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1011ec0) 00:21:57.183 [2024-07-15 18:07:50.900784] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.183 [2024-07-15 18:07:50.900793] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x10952c0, cid 3, qid 0 00:21:57.183 [2024-07-15 18:07:50.900868] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.183 [2024-07-15 18:07:50.900873] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.183 [2024-07-15 18:07:50.900878] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.900882] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x10952c0) on tqpair=0x1011ec0 00:21:57.183 [2024-07-15 18:07:50.900890] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.900894] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.900897] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1011ec0) 00:21:57.183 [2024-07-15 18:07:50.900903] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.183 [2024-07-15 18:07:50.900912] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x10952c0, cid 3, qid 0 00:21:57.183 [2024-07-15 18:07:50.900984] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.183 [2024-07-15 18:07:50.900990] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.183 [2024-07-15 18:07:50.900993] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.900996] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x10952c0) on tqpair=0x1011ec0 00:21:57.183 [2024-07-15 18:07:50.901004] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.901008] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.901011] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1011ec0) 00:21:57.183 [2024-07-15 18:07:50.901017] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.183 [2024-07-15 18:07:50.901026] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x10952c0, cid 3, qid 0 00:21:57.183 [2024-07-15 18:07:50.901103] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.183 [2024-07-15 18:07:50.901109] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.183 [2024-07-15 18:07:50.901112] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.901115] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x10952c0) on tqpair=0x1011ec0 00:21:57.183 [2024-07-15 18:07:50.901123] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.901127] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.901130] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1011ec0) 00:21:57.183 [2024-07-15 18:07:50.901136] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.183 [2024-07-15 18:07:50.901145] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x10952c0, cid 3, qid 0 00:21:57.183 [2024-07-15 18:07:50.901219] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.183 [2024-07-15 18:07:50.901230] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.183 [2024-07-15 18:07:50.901233] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.901236] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x10952c0) on tqpair=0x1011ec0 00:21:57.183 [2024-07-15 18:07:50.901244] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.901248] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.901251] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1011ec0) 00:21:57.183 [2024-07-15 18:07:50.901257] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.183 [2024-07-15 18:07:50.901267] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x10952c0, cid 3, qid 0 00:21:57.183 [2024-07-15 18:07:50.901337] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.183 [2024-07-15 18:07:50.901343] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.183 [2024-07-15 18:07:50.901346] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.901351] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x10952c0) on tqpair=0x1011ec0 00:21:57.183 [2024-07-15 18:07:50.901359] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.901363] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.901366] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1011ec0) 00:21:57.183 [2024-07-15 18:07:50.901372] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.183 [2024-07-15 18:07:50.901381] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x10952c0, cid 3, qid 0 00:21:57.183 [2024-07-15 18:07:50.901456] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.183 [2024-07-15 18:07:50.901462] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.183 [2024-07-15 18:07:50.901465] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.901468] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x10952c0) on tqpair=0x1011ec0 00:21:57.183 [2024-07-15 18:07:50.901476] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.901480] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.901483] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1011ec0) 00:21:57.183 [2024-07-15 18:07:50.901489] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.183 [2024-07-15 18:07:50.901498] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x10952c0, cid 3, qid 0 00:21:57.183 [2024-07-15 18:07:50.901572] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.183 [2024-07-15 18:07:50.901578] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.183 [2024-07-15 18:07:50.901581] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.901584] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x10952c0) on tqpair=0x1011ec0 00:21:57.183 [2024-07-15 18:07:50.901592] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.901596] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.183 [2024-07-15 18:07:50.901599] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1011ec0) 00:21:57.184 [2024-07-15 18:07:50.901604] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.184 [2024-07-15 18:07:50.901614] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x10952c0, cid 3, qid 0 00:21:57.184 [2024-07-15 18:07:50.901691] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.184 [2024-07-15 18:07:50.901697] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.184 [2024-07-15 18:07:50.901700] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.184 [2024-07-15 18:07:50.901703] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x10952c0) on tqpair=0x1011ec0 00:21:57.184 [2024-07-15 18:07:50.901711] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.184 [2024-07-15 18:07:50.901715] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.184 [2024-07-15 18:07:50.901718] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1011ec0) 00:21:57.184 [2024-07-15 18:07:50.901723] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.184 [2024-07-15 18:07:50.901733] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x10952c0, cid 3, qid 0 00:21:57.184 [2024-07-15 18:07:50.901805] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.184 [2024-07-15 18:07:50.901810] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.184 [2024-07-15 18:07:50.901813] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.184 [2024-07-15 18:07:50.901817] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x10952c0) on tqpair=0x1011ec0 00:21:57.184 [2024-07-15 18:07:50.901827] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.184 [2024-07-15 18:07:50.901830] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.184 [2024-07-15 18:07:50.901833] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1011ec0) 00:21:57.184 [2024-07-15 18:07:50.901839] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.184 [2024-07-15 18:07:50.901848] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x10952c0, cid 3, qid 0 00:21:57.184 [2024-07-15 18:07:50.901925] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.184 [2024-07-15 18:07:50.901931] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.184 [2024-07-15 18:07:50.901934] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.184 [2024-07-15 18:07:50.901937] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x10952c0) on tqpair=0x1011ec0 00:21:57.184 [2024-07-15 18:07:50.901945] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.184 [2024-07-15 18:07:50.901949] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.184 [2024-07-15 18:07:50.901952] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1011ec0) 00:21:57.184 [2024-07-15 18:07:50.901958] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.184 [2024-07-15 18:07:50.901967] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x10952c0, cid 3, qid 0 00:21:57.184 [2024-07-15 18:07:50.902043] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.184 [2024-07-15 18:07:50.902049] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.184 [2024-07-15 18:07:50.902052] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.184 [2024-07-15 18:07:50.902055] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x10952c0) on tqpair=0x1011ec0 00:21:57.184 [2024-07-15 18:07:50.902063] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.184 [2024-07-15 18:07:50.902067] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.184 [2024-07-15 18:07:50.902070] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1011ec0) 00:21:57.184 [2024-07-15 18:07:50.902075] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.184 [2024-07-15 18:07:50.902085] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x10952c0, cid 3, qid 0 00:21:57.184 [2024-07-15 18:07:50.902159] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.184 [2024-07-15 18:07:50.902164] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.184 [2024-07-15 18:07:50.902167] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.184 [2024-07-15 18:07:50.902170] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x10952c0) on tqpair=0x1011ec0 00:21:57.184 [2024-07-15 18:07:50.902179] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.184 [2024-07-15 18:07:50.902182] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.184 [2024-07-15 18:07:50.902185] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1011ec0) 00:21:57.184 [2024-07-15 18:07:50.902191] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.184 [2024-07-15 18:07:50.902200] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x10952c0, cid 3, qid 0 00:21:57.447 [2024-07-15 18:07:50.906231] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.447 [2024-07-15 18:07:50.906241] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.447 [2024-07-15 18:07:50.906244] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.447 [2024-07-15 18:07:50.906247] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x10952c0) on tqpair=0x1011ec0 00:21:57.447 [2024-07-15 18:07:50.906257] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.447 [2024-07-15 18:07:50.906267] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.447 [2024-07-15 18:07:50.906270] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x1011ec0) 00:21:57.447 [2024-07-15 18:07:50.906276] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.447 [2024-07-15 18:07:50.906288] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x10952c0, cid 3, qid 0 00:21:57.447 [2024-07-15 18:07:50.906447] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.447 [2024-07-15 18:07:50.906452] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.447 [2024-07-15 18:07:50.906455] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.447 [2024-07-15 18:07:50.906458] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x10952c0) on tqpair=0x1011ec0 00:21:57.447 [2024-07-15 18:07:50.906465] nvme_ctrlr.c:1269:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [nqn.2014-08.org.nvmexpress.discovery] shutdown complete in 5 milliseconds 00:21:57.447 00:21:57.447 18:07:50 nvmf_tcp.nvmf_identify -- host/identify.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' -L all 00:21:57.447 [2024-07-15 18:07:50.942964] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:21:57.447 [2024-07-15 18:07:50.943001] [ DPDK EAL parameters: identify --no-shconf -c 0x1 -n 1 -m 0 --no-pci --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid668124 ] 00:21:57.447 EAL: No free 2048 kB hugepages reported on node 1 00:21:57.447 [2024-07-15 18:07:50.970454] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to connect adminq (no timeout) 00:21:57.447 [2024-07-15 18:07:50.970496] nvme_tcp.c:2338:nvme_tcp_qpair_connect_sock: *DEBUG*: adrfam 1 ai_family 2 00:21:57.447 [2024-07-15 18:07:50.970500] nvme_tcp.c:2342:nvme_tcp_qpair_connect_sock: *DEBUG*: trsvcid is 4420 00:21:57.447 [2024-07-15 18:07:50.970513] nvme_tcp.c:2360:nvme_tcp_qpair_connect_sock: *DEBUG*: sock_impl_name is (null) 00:21:57.447 [2024-07-15 18:07:50.970519] sock.c: 337:spdk_sock_connect_ext: *DEBUG*: Creating a client socket using impl posix 00:21:57.447 [2024-07-15 18:07:50.970822] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for connect adminq (no timeout) 00:21:57.447 [2024-07-15 18:07:50.970845] nvme_tcp.c:1555:nvme_tcp_send_icreq_complete: *DEBUG*: Complete the icreq send for tqpair=0x7a9ec0 0 00:21:57.447 [2024-07-15 18:07:50.984232] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 1 00:21:57.447 [2024-07-15 18:07:50.984243] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =1 00:21:57.447 [2024-07-15 18:07:50.984247] nvme_tcp.c:1601:nvme_tcp_icresp_handle: *DEBUG*: host_hdgst_enable: 0 00:21:57.447 [2024-07-15 18:07:50.984249] nvme_tcp.c:1602:nvme_tcp_icresp_handle: *DEBUG*: host_ddgst_enable: 0 00:21:57.447 [2024-07-15 18:07:50.984276] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.447 [2024-07-15 18:07:50.984280] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.447 [2024-07-15 18:07:50.984284] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x7a9ec0) 00:21:57.447 [2024-07-15 18:07:50.984293] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC CONNECT qid:0 cid:0 SGL DATA BLOCK OFFSET 0x0 len:0x400 00:21:57.447 [2024-07-15 18:07:50.984308] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82ce40, cid 0, qid 0 00:21:57.447 [2024-07-15 18:07:50.991233] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.447 [2024-07-15 18:07:50.991241] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.447 [2024-07-15 18:07:50.991244] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.447 [2024-07-15 18:07:50.991251] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82ce40) on tqpair=0x7a9ec0 00:21:57.447 [2024-07-15 18:07:50.991259] nvme_fabric.c: 622:_nvme_fabric_qpair_connect_poll: *DEBUG*: CNTLID 0x0001 00:21:57.447 [2024-07-15 18:07:50.991264] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to read vs (no timeout) 00:21:57.447 [2024-07-15 18:07:50.991269] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to read vs wait for vs (no timeout) 00:21:57.447 [2024-07-15 18:07:50.991278] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.447 [2024-07-15 18:07:50.991282] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.447 [2024-07-15 18:07:50.991285] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x7a9ec0) 00:21:57.447 [2024-07-15 18:07:50.991292] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.447 [2024-07-15 18:07:50.991305] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82ce40, cid 0, qid 0 00:21:57.447 [2024-07-15 18:07:50.991463] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.448 [2024-07-15 18:07:50.991469] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.448 [2024-07-15 18:07:50.991472] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.991475] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82ce40) on tqpair=0x7a9ec0 00:21:57.448 [2024-07-15 18:07:50.991479] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to read cap (no timeout) 00:21:57.448 [2024-07-15 18:07:50.991486] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to read cap wait for cap (no timeout) 00:21:57.448 [2024-07-15 18:07:50.991492] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.991495] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.991498] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x7a9ec0) 00:21:57.448 [2024-07-15 18:07:50.991504] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.448 [2024-07-15 18:07:50.991514] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82ce40, cid 0, qid 0 00:21:57.448 [2024-07-15 18:07:50.991588] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.448 [2024-07-15 18:07:50.991593] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.448 [2024-07-15 18:07:50.991597] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.991600] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82ce40) on tqpair=0x7a9ec0 00:21:57.448 [2024-07-15 18:07:50.991604] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to check en (no timeout) 00:21:57.448 [2024-07-15 18:07:50.991610] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to check en wait for cc (timeout 15000 ms) 00:21:57.448 [2024-07-15 18:07:50.991616] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.991620] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.991623] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x7a9ec0) 00:21:57.448 [2024-07-15 18:07:50.991629] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.448 [2024-07-15 18:07:50.991638] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82ce40, cid 0, qid 0 00:21:57.448 [2024-07-15 18:07:50.991707] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.448 [2024-07-15 18:07:50.991712] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.448 [2024-07-15 18:07:50.991715] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.991719] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82ce40) on tqpair=0x7a9ec0 00:21:57.448 [2024-07-15 18:07:50.991725] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to disable and wait for CSTS.RDY = 0 (timeout 15000 ms) 00:21:57.448 [2024-07-15 18:07:50.991733] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.991736] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.991739] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x7a9ec0) 00:21:57.448 [2024-07-15 18:07:50.991745] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.448 [2024-07-15 18:07:50.991754] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82ce40, cid 0, qid 0 00:21:57.448 [2024-07-15 18:07:50.991827] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.448 [2024-07-15 18:07:50.991833] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.448 [2024-07-15 18:07:50.991836] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.991839] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82ce40) on tqpair=0x7a9ec0 00:21:57.448 [2024-07-15 18:07:50.991843] nvme_ctrlr.c:3869:nvme_ctrlr_process_init_wait_for_ready_0: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] CC.EN = 0 && CSTS.RDY = 0 00:21:57.448 [2024-07-15 18:07:50.991847] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to controller is disabled (timeout 15000 ms) 00:21:57.448 [2024-07-15 18:07:50.991853] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to enable controller by writing CC.EN = 1 (timeout 15000 ms) 00:21:57.448 [2024-07-15 18:07:50.991958] nvme_ctrlr.c:4062:nvme_ctrlr_process_init: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] Setting CC.EN = 1 00:21:57.448 [2024-07-15 18:07:50.991961] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to enable controller by writing CC.EN = 1 reg (timeout 15000 ms) 00:21:57.448 [2024-07-15 18:07:50.991967] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.991971] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.991974] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x7a9ec0) 00:21:57.448 [2024-07-15 18:07:50.991980] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.448 [2024-07-15 18:07:50.991989] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82ce40, cid 0, qid 0 00:21:57.448 [2024-07-15 18:07:50.992064] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.448 [2024-07-15 18:07:50.992069] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.448 [2024-07-15 18:07:50.992072] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.992075] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82ce40) on tqpair=0x7a9ec0 00:21:57.448 [2024-07-15 18:07:50.992079] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for CSTS.RDY = 1 (timeout 15000 ms) 00:21:57.448 [2024-07-15 18:07:50.992087] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.992090] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.992094] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x7a9ec0) 00:21:57.448 [2024-07-15 18:07:50.992099] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:0 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.448 [2024-07-15 18:07:50.992108] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82ce40, cid 0, qid 0 00:21:57.448 [2024-07-15 18:07:50.992180] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.448 [2024-07-15 18:07:50.992186] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.448 [2024-07-15 18:07:50.992189] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.992193] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82ce40) on tqpair=0x7a9ec0 00:21:57.448 [2024-07-15 18:07:50.992197] nvme_ctrlr.c:3904:nvme_ctrlr_process_init_enable_wait_for_ready_1: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] CC.EN = 1 && CSTS.RDY = 1 - controller is ready 00:21:57.448 [2024-07-15 18:07:50.992201] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to reset admin queue (timeout 30000 ms) 00:21:57.448 [2024-07-15 18:07:50.992207] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to identify controller (no timeout) 00:21:57.448 [2024-07-15 18:07:50.992214] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for identify controller (timeout 30000 ms) 00:21:57.448 [2024-07-15 18:07:50.992221] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.992229] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x7a9ec0) 00:21:57.448 [2024-07-15 18:07:50.992235] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:0 nsid:0 cdw10:00000001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.448 [2024-07-15 18:07:50.992245] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82ce40, cid 0, qid 0 00:21:57.448 [2024-07-15 18:07:50.992357] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:57.448 [2024-07-15 18:07:50.992363] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:57.448 [2024-07-15 18:07:50.992366] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.992369] nvme_tcp.c:1720:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x7a9ec0): datao=0, datal=4096, cccid=0 00:21:57.448 [2024-07-15 18:07:50.992373] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x82ce40) on tqpair(0x7a9ec0): expected_datao=0, payload_size=4096 00:21:57.448 [2024-07-15 18:07:50.992377] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.992383] nvme_tcp.c:1521:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.992386] nvme_tcp.c:1312:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.992408] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.448 [2024-07-15 18:07:50.992413] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.448 [2024-07-15 18:07:50.992416] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.992419] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82ce40) on tqpair=0x7a9ec0 00:21:57.448 [2024-07-15 18:07:50.992425] nvme_ctrlr.c:2053:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] transport max_xfer_size 4294967295 00:21:57.448 [2024-07-15 18:07:50.992431] nvme_ctrlr.c:2057:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] MDTS max_xfer_size 131072 00:21:57.448 [2024-07-15 18:07:50.992436] nvme_ctrlr.c:2060:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] CNTLID 0x0001 00:21:57.448 [2024-07-15 18:07:50.992439] nvme_ctrlr.c:2084:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] transport max_sges 16 00:21:57.448 [2024-07-15 18:07:50.992443] nvme_ctrlr.c:2099:nvme_ctrlr_identify_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] fuses compare and write: 1 00:21:57.448 [2024-07-15 18:07:50.992447] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to configure AER (timeout 30000 ms) 00:21:57.448 [2024-07-15 18:07:50.992454] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for configure aer (timeout 30000 ms) 00:21:57.448 [2024-07-15 18:07:50.992460] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.992463] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.992466] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x7a9ec0) 00:21:57.448 [2024-07-15 18:07:50.992472] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES ASYNC EVENT CONFIGURATION cid:0 cdw10:0000000b SGL DATA BLOCK OFFSET 0x0 len:0x0 00:21:57.448 [2024-07-15 18:07:50.992484] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82ce40, cid 0, qid 0 00:21:57.448 [2024-07-15 18:07:50.992560] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.448 [2024-07-15 18:07:50.992566] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.448 [2024-07-15 18:07:50.992569] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.992572] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82ce40) on tqpair=0x7a9ec0 00:21:57.448 [2024-07-15 18:07:50.992578] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.992581] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.992584] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=0 on tqpair(0x7a9ec0) 00:21:57.448 [2024-07-15 18:07:50.992589] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:21:57.448 [2024-07-15 18:07:50.992594] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.992597] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.992600] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=1 on tqpair(0x7a9ec0) 00:21:57.448 [2024-07-15 18:07:50.992605] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:21:57.448 [2024-07-15 18:07:50.992611] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.992614] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.448 [2024-07-15 18:07:50.992617] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=2 on tqpair(0x7a9ec0) 00:21:57.449 [2024-07-15 18:07:50.992622] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:21:57.449 [2024-07-15 18:07:50.992626] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:50.992630] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:50.992633] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.449 [2024-07-15 18:07:50.992638] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:21:57.449 [2024-07-15 18:07:50.992642] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to set keep alive timeout (timeout 30000 ms) 00:21:57.449 [2024-07-15 18:07:50.992651] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for set keep alive timeout (timeout 30000 ms) 00:21:57.449 [2024-07-15 18:07:50.992657] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:50.992660] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x7a9ec0) 00:21:57.449 [2024-07-15 18:07:50.992666] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES KEEP ALIVE TIMER cid:4 cdw10:0000000f SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.449 [2024-07-15 18:07:50.992676] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82ce40, cid 0, qid 0 00:21:57.449 [2024-07-15 18:07:50.992680] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82cfc0, cid 1, qid 0 00:21:57.449 [2024-07-15 18:07:50.992684] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d140, cid 2, qid 0 00:21:57.449 [2024-07-15 18:07:50.992688] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.449 [2024-07-15 18:07:50.992692] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d440, cid 4, qid 0 00:21:57.449 [2024-07-15 18:07:50.992797] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.449 [2024-07-15 18:07:50.992803] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.449 [2024-07-15 18:07:50.992806] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:50.992809] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d440) on tqpair=0x7a9ec0 00:21:57.449 [2024-07-15 18:07:50.992815] nvme_ctrlr.c:3022:nvme_ctrlr_set_keep_alive_timeout_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] Sending keep alive every 5000000 us 00:21:57.449 [2024-07-15 18:07:50.992819] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to identify controller iocs specific (timeout 30000 ms) 00:21:57.449 [2024-07-15 18:07:50.992825] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to set number of queues (timeout 30000 ms) 00:21:57.449 [2024-07-15 18:07:50.992831] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for set number of queues (timeout 30000 ms) 00:21:57.449 [2024-07-15 18:07:50.992836] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:50.992840] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:50.992843] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x7a9ec0) 00:21:57.449 [2024-07-15 18:07:50.992848] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: SET FEATURES NUMBER OF QUEUES cid:4 cdw10:00000007 SGL DATA BLOCK OFFSET 0x0 len:0x0 00:21:57.449 [2024-07-15 18:07:50.992857] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d440, cid 4, qid 0 00:21:57.449 [2024-07-15 18:07:50.992931] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.449 [2024-07-15 18:07:50.992937] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.449 [2024-07-15 18:07:50.992939] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:50.992942] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d440) on tqpair=0x7a9ec0 00:21:57.449 [2024-07-15 18:07:50.992992] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to identify active ns (timeout 30000 ms) 00:21:57.449 [2024-07-15 18:07:50.993001] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for identify active ns (timeout 30000 ms) 00:21:57.449 [2024-07-15 18:07:50.993007] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:50.993011] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x7a9ec0) 00:21:57.449 [2024-07-15 18:07:50.993016] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:0 cdw10:00000002 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.449 [2024-07-15 18:07:50.993025] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d440, cid 4, qid 0 00:21:57.449 [2024-07-15 18:07:50.993122] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:57.449 [2024-07-15 18:07:50.993128] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:57.449 [2024-07-15 18:07:50.993131] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:50.993134] nvme_tcp.c:1720:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x7a9ec0): datao=0, datal=4096, cccid=4 00:21:57.449 [2024-07-15 18:07:50.993138] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x82d440) on tqpair(0x7a9ec0): expected_datao=0, payload_size=4096 00:21:57.449 [2024-07-15 18:07:50.993142] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:50.993147] nvme_tcp.c:1521:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:50.993151] nvme_tcp.c:1312:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:51.038231] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.449 [2024-07-15 18:07:51.038242] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.449 [2024-07-15 18:07:51.038246] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:51.038249] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d440) on tqpair=0x7a9ec0 00:21:57.449 [2024-07-15 18:07:51.038258] nvme_ctrlr.c:4693:spdk_nvme_ctrlr_get_ns: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] Namespace 1 was added 00:21:57.449 [2024-07-15 18:07:51.038271] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to identify ns (timeout 30000 ms) 00:21:57.449 [2024-07-15 18:07:51.038282] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for identify ns (timeout 30000 ms) 00:21:57.449 [2024-07-15 18:07:51.038289] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:51.038293] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x7a9ec0) 00:21:57.449 [2024-07-15 18:07:51.038300] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:1 cdw10:00000000 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.449 [2024-07-15 18:07:51.038313] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d440, cid 4, qid 0 00:21:57.449 [2024-07-15 18:07:51.038485] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:57.449 [2024-07-15 18:07:51.038491] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:57.449 [2024-07-15 18:07:51.038495] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:51.038498] nvme_tcp.c:1720:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x7a9ec0): datao=0, datal=4096, cccid=4 00:21:57.449 [2024-07-15 18:07:51.038501] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x82d440) on tqpair(0x7a9ec0): expected_datao=0, payload_size=4096 00:21:57.449 [2024-07-15 18:07:51.038505] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:51.038529] nvme_tcp.c:1521:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:51.038533] nvme_tcp.c:1312:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:51.079363] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.449 [2024-07-15 18:07:51.079375] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.449 [2024-07-15 18:07:51.079378] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:51.079382] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d440) on tqpair=0x7a9ec0 00:21:57.449 [2024-07-15 18:07:51.079394] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to identify namespace id descriptors (timeout 30000 ms) 00:21:57.449 [2024-07-15 18:07:51.079404] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to wait for identify namespace id descriptors (timeout 30000 ms) 00:21:57.449 [2024-07-15 18:07:51.079412] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:51.079415] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x7a9ec0) 00:21:57.449 [2024-07-15 18:07:51.079422] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: IDENTIFY (06) qid:0 cid:4 nsid:1 cdw10:00000003 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.449 [2024-07-15 18:07:51.079433] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d440, cid 4, qid 0 00:21:57.449 [2024-07-15 18:07:51.079513] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:57.449 [2024-07-15 18:07:51.079519] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:57.449 [2024-07-15 18:07:51.079522] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:51.079525] nvme_tcp.c:1720:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x7a9ec0): datao=0, datal=4096, cccid=4 00:21:57.449 [2024-07-15 18:07:51.079529] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x82d440) on tqpair(0x7a9ec0): expected_datao=0, payload_size=4096 00:21:57.449 [2024-07-15 18:07:51.079533] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:51.079556] nvme_tcp.c:1521:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:51.079560] nvme_tcp.c:1312:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:51.120371] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.449 [2024-07-15 18:07:51.120382] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.449 [2024-07-15 18:07:51.120386] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:51.120389] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d440) on tqpair=0x7a9ec0 00:21:57.449 [2024-07-15 18:07:51.120403] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to identify ns iocs specific (timeout 30000 ms) 00:21:57.449 [2024-07-15 18:07:51.120412] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to set supported log pages (timeout 30000 ms) 00:21:57.449 [2024-07-15 18:07:51.120419] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to set supported features (timeout 30000 ms) 00:21:57.449 [2024-07-15 18:07:51.120425] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to set host behavior support feature (timeout 30000 ms) 00:21:57.449 [2024-07-15 18:07:51.120430] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to set doorbell buffer config (timeout 30000 ms) 00:21:57.449 [2024-07-15 18:07:51.120434] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to set host ID (timeout 30000 ms) 00:21:57.449 [2024-07-15 18:07:51.120439] nvme_ctrlr.c:3110:nvme_ctrlr_set_host_id: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] NVMe-oF transport - not sending Set Features - Host ID 00:21:57.449 [2024-07-15 18:07:51.120443] nvme_ctrlr.c:1553:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to transport ready (timeout 30000 ms) 00:21:57.449 [2024-07-15 18:07:51.120447] nvme_ctrlr.c:1559:_nvme_ctrlr_set_state: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] setting state to ready (no timeout) 00:21:57.449 [2024-07-15 18:07:51.120459] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:51.120463] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x7a9ec0) 00:21:57.449 [2024-07-15 18:07:51.120469] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ARBITRATION cid:4 cdw10:00000001 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.449 [2024-07-15 18:07:51.120475] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:51.120479] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.449 [2024-07-15 18:07:51.120481] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x7a9ec0) 00:21:57.449 [2024-07-15 18:07:51.120487] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:5 nsid:0 cdw10:00000000 cdw11:00000000 00:21:57.449 [2024-07-15 18:07:51.120500] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d440, cid 4, qid 0 00:21:57.450 [2024-07-15 18:07:51.120505] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d5c0, cid 5, qid 0 00:21:57.450 [2024-07-15 18:07:51.120600] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.450 [2024-07-15 18:07:51.120606] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.450 [2024-07-15 18:07:51.120609] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.120612] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d440) on tqpair=0x7a9ec0 00:21:57.450 [2024-07-15 18:07:51.120618] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.450 [2024-07-15 18:07:51.120623] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.450 [2024-07-15 18:07:51.120626] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.120629] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d5c0) on tqpair=0x7a9ec0 00:21:57.450 [2024-07-15 18:07:51.120638] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.120641] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x7a9ec0) 00:21:57.450 [2024-07-15 18:07:51.120647] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES POWER MANAGEMENT cid:5 cdw10:00000002 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.450 [2024-07-15 18:07:51.120656] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d5c0, cid 5, qid 0 00:21:57.450 [2024-07-15 18:07:51.120731] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.450 [2024-07-15 18:07:51.120739] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.450 [2024-07-15 18:07:51.120742] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.120745] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d5c0) on tqpair=0x7a9ec0 00:21:57.450 [2024-07-15 18:07:51.120753] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.120757] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x7a9ec0) 00:21:57.450 [2024-07-15 18:07:51.120762] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES TEMPERATURE THRESHOLD cid:5 cdw10:00000004 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.450 [2024-07-15 18:07:51.120771] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d5c0, cid 5, qid 0 00:21:57.450 [2024-07-15 18:07:51.120849] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.450 [2024-07-15 18:07:51.120855] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.450 [2024-07-15 18:07:51.120858] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.120861] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d5c0) on tqpair=0x7a9ec0 00:21:57.450 [2024-07-15 18:07:51.120869] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.120872] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x7a9ec0) 00:21:57.450 [2024-07-15 18:07:51.120878] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES NUMBER OF QUEUES cid:5 cdw10:00000007 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.450 [2024-07-15 18:07:51.120886] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d5c0, cid 5, qid 0 00:21:57.450 [2024-07-15 18:07:51.120965] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.450 [2024-07-15 18:07:51.120971] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.450 [2024-07-15 18:07:51.120974] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.120977] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d5c0) on tqpair=0x7a9ec0 00:21:57.450 [2024-07-15 18:07:51.120990] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.120994] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=5 on tqpair(0x7a9ec0) 00:21:57.450 [2024-07-15 18:07:51.121000] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:5 nsid:ffffffff cdw10:07ff0001 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.450 [2024-07-15 18:07:51.121006] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.121009] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=4 on tqpair(0x7a9ec0) 00:21:57.450 [2024-07-15 18:07:51.121014] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:4 nsid:ffffffff cdw10:007f0002 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.450 [2024-07-15 18:07:51.121021] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.121024] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=6 on tqpair(0x7a9ec0) 00:21:57.450 [2024-07-15 18:07:51.121029] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:6 nsid:ffffffff cdw10:007f0003 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.450 [2024-07-15 18:07:51.121035] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.121038] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=7 on tqpair(0x7a9ec0) 00:21:57.450 [2024-07-15 18:07:51.121044] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: GET LOG PAGE (02) qid:0 cid:7 nsid:ffffffff cdw10:03ff0005 cdw11:00000000 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.450 [2024-07-15 18:07:51.121054] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d5c0, cid 5, qid 0 00:21:57.450 [2024-07-15 18:07:51.121058] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d440, cid 4, qid 0 00:21:57.450 [2024-07-15 18:07:51.121062] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d740, cid 6, qid 0 00:21:57.450 [2024-07-15 18:07:51.121068] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d8c0, cid 7, qid 0 00:21:57.450 [2024-07-15 18:07:51.125237] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:57.450 [2024-07-15 18:07:51.125247] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:57.450 [2024-07-15 18:07:51.125250] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.125253] nvme_tcp.c:1720:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x7a9ec0): datao=0, datal=8192, cccid=5 00:21:57.450 [2024-07-15 18:07:51.125257] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x82d5c0) on tqpair(0x7a9ec0): expected_datao=0, payload_size=8192 00:21:57.450 [2024-07-15 18:07:51.125261] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.125267] nvme_tcp.c:1521:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.125270] nvme_tcp.c:1312:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.125275] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:57.450 [2024-07-15 18:07:51.125280] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:57.450 [2024-07-15 18:07:51.125283] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.125286] nvme_tcp.c:1720:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x7a9ec0): datao=0, datal=512, cccid=4 00:21:57.450 [2024-07-15 18:07:51.125290] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x82d440) on tqpair(0x7a9ec0): expected_datao=0, payload_size=512 00:21:57.450 [2024-07-15 18:07:51.125294] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.125299] nvme_tcp.c:1521:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.125302] nvme_tcp.c:1312:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.125307] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:57.450 [2024-07-15 18:07:51.125311] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:57.450 [2024-07-15 18:07:51.125314] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.125317] nvme_tcp.c:1720:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x7a9ec0): datao=0, datal=512, cccid=6 00:21:57.450 [2024-07-15 18:07:51.125321] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x82d740) on tqpair(0x7a9ec0): expected_datao=0, payload_size=512 00:21:57.450 [2024-07-15 18:07:51.125325] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.125330] nvme_tcp.c:1521:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.125333] nvme_tcp.c:1312:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.125338] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 7 00:21:57.450 [2024-07-15 18:07:51.125342] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =7 00:21:57.450 [2024-07-15 18:07:51.125345] nvme_tcp.c:1719:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.125349] nvme_tcp.c:1720:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: c2h_data info on tqpair(0x7a9ec0): datao=0, datal=4096, cccid=7 00:21:57.450 [2024-07-15 18:07:51.125352] nvme_tcp.c:1731:nvme_tcp_c2h_data_hdr_handle: *DEBUG*: tcp_req(0x82d8c0) on tqpair(0x7a9ec0): expected_datao=0, payload_size=4096 00:21:57.450 [2024-07-15 18:07:51.125356] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.125361] nvme_tcp.c:1521:nvme_tcp_pdu_payload_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.125364] nvme_tcp.c:1312:nvme_tcp_c2h_data_payload_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.125369] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.450 [2024-07-15 18:07:51.125374] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.450 [2024-07-15 18:07:51.125377] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.125380] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d5c0) on tqpair=0x7a9ec0 00:21:57.450 [2024-07-15 18:07:51.125391] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.450 [2024-07-15 18:07:51.125399] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.450 [2024-07-15 18:07:51.125402] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.125406] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d440) on tqpair=0x7a9ec0 00:21:57.450 [2024-07-15 18:07:51.125414] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.450 [2024-07-15 18:07:51.125419] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.450 [2024-07-15 18:07:51.125422] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.125425] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d740) on tqpair=0x7a9ec0 00:21:57.450 [2024-07-15 18:07:51.125431] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.450 [2024-07-15 18:07:51.125436] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.450 [2024-07-15 18:07:51.125439] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.450 [2024-07-15 18:07:51.125442] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d8c0) on tqpair=0x7a9ec0 00:21:57.450 ===================================================== 00:21:57.450 NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:21:57.450 ===================================================== 00:21:57.450 Controller Capabilities/Features 00:21:57.450 ================================ 00:21:57.450 Vendor ID: 8086 00:21:57.450 Subsystem Vendor ID: 8086 00:21:57.450 Serial Number: SPDK00000000000001 00:21:57.450 Model Number: SPDK bdev Controller 00:21:57.450 Firmware Version: 24.09 00:21:57.450 Recommended Arb Burst: 6 00:21:57.450 IEEE OUI Identifier: e4 d2 5c 00:21:57.450 Multi-path I/O 00:21:57.450 May have multiple subsystem ports: Yes 00:21:57.450 May have multiple controllers: Yes 00:21:57.450 Associated with SR-IOV VF: No 00:21:57.450 Max Data Transfer Size: 131072 00:21:57.450 Max Number of Namespaces: 32 00:21:57.450 Max Number of I/O Queues: 127 00:21:57.450 NVMe Specification Version (VS): 1.3 00:21:57.450 NVMe Specification Version (Identify): 1.3 00:21:57.450 Maximum Queue Entries: 128 00:21:57.450 Contiguous Queues Required: Yes 00:21:57.450 Arbitration Mechanisms Supported 00:21:57.450 Weighted Round Robin: Not Supported 00:21:57.450 Vendor Specific: Not Supported 00:21:57.450 Reset Timeout: 15000 ms 00:21:57.450 Doorbell Stride: 4 bytes 00:21:57.450 NVM Subsystem Reset: Not Supported 00:21:57.450 Command Sets Supported 00:21:57.451 NVM Command Set: Supported 00:21:57.451 Boot Partition: Not Supported 00:21:57.451 Memory Page Size Minimum: 4096 bytes 00:21:57.451 Memory Page Size Maximum: 4096 bytes 00:21:57.451 Persistent Memory Region: Not Supported 00:21:57.451 Optional Asynchronous Events Supported 00:21:57.451 Namespace Attribute Notices: Supported 00:21:57.451 Firmware Activation Notices: Not Supported 00:21:57.451 ANA Change Notices: Not Supported 00:21:57.451 PLE Aggregate Log Change Notices: Not Supported 00:21:57.451 LBA Status Info Alert Notices: Not Supported 00:21:57.451 EGE Aggregate Log Change Notices: Not Supported 00:21:57.451 Normal NVM Subsystem Shutdown event: Not Supported 00:21:57.451 Zone Descriptor Change Notices: Not Supported 00:21:57.451 Discovery Log Change Notices: Not Supported 00:21:57.451 Controller Attributes 00:21:57.451 128-bit Host Identifier: Supported 00:21:57.451 Non-Operational Permissive Mode: Not Supported 00:21:57.451 NVM Sets: Not Supported 00:21:57.451 Read Recovery Levels: Not Supported 00:21:57.451 Endurance Groups: Not Supported 00:21:57.451 Predictable Latency Mode: Not Supported 00:21:57.451 Traffic Based Keep ALive: Not Supported 00:21:57.451 Namespace Granularity: Not Supported 00:21:57.451 SQ Associations: Not Supported 00:21:57.451 UUID List: Not Supported 00:21:57.451 Multi-Domain Subsystem: Not Supported 00:21:57.451 Fixed Capacity Management: Not Supported 00:21:57.451 Variable Capacity Management: Not Supported 00:21:57.451 Delete Endurance Group: Not Supported 00:21:57.451 Delete NVM Set: Not Supported 00:21:57.451 Extended LBA Formats Supported: Not Supported 00:21:57.451 Flexible Data Placement Supported: Not Supported 00:21:57.451 00:21:57.451 Controller Memory Buffer Support 00:21:57.451 ================================ 00:21:57.451 Supported: No 00:21:57.451 00:21:57.451 Persistent Memory Region Support 00:21:57.451 ================================ 00:21:57.451 Supported: No 00:21:57.451 00:21:57.451 Admin Command Set Attributes 00:21:57.451 ============================ 00:21:57.451 Security Send/Receive: Not Supported 00:21:57.451 Format NVM: Not Supported 00:21:57.451 Firmware Activate/Download: Not Supported 00:21:57.451 Namespace Management: Not Supported 00:21:57.451 Device Self-Test: Not Supported 00:21:57.451 Directives: Not Supported 00:21:57.451 NVMe-MI: Not Supported 00:21:57.451 Virtualization Management: Not Supported 00:21:57.451 Doorbell Buffer Config: Not Supported 00:21:57.451 Get LBA Status Capability: Not Supported 00:21:57.451 Command & Feature Lockdown Capability: Not Supported 00:21:57.451 Abort Command Limit: 4 00:21:57.451 Async Event Request Limit: 4 00:21:57.451 Number of Firmware Slots: N/A 00:21:57.451 Firmware Slot 1 Read-Only: N/A 00:21:57.451 Firmware Activation Without Reset: N/A 00:21:57.451 Multiple Update Detection Support: N/A 00:21:57.451 Firmware Update Granularity: No Information Provided 00:21:57.451 Per-Namespace SMART Log: No 00:21:57.451 Asymmetric Namespace Access Log Page: Not Supported 00:21:57.451 Subsystem NQN: nqn.2016-06.io.spdk:cnode1 00:21:57.451 Command Effects Log Page: Supported 00:21:57.451 Get Log Page Extended Data: Supported 00:21:57.451 Telemetry Log Pages: Not Supported 00:21:57.451 Persistent Event Log Pages: Not Supported 00:21:57.451 Supported Log Pages Log Page: May Support 00:21:57.451 Commands Supported & Effects Log Page: Not Supported 00:21:57.451 Feature Identifiers & Effects Log Page:May Support 00:21:57.451 NVMe-MI Commands & Effects Log Page: May Support 00:21:57.451 Data Area 4 for Telemetry Log: Not Supported 00:21:57.451 Error Log Page Entries Supported: 128 00:21:57.451 Keep Alive: Supported 00:21:57.451 Keep Alive Granularity: 10000 ms 00:21:57.451 00:21:57.451 NVM Command Set Attributes 00:21:57.451 ========================== 00:21:57.451 Submission Queue Entry Size 00:21:57.451 Max: 64 00:21:57.451 Min: 64 00:21:57.451 Completion Queue Entry Size 00:21:57.451 Max: 16 00:21:57.451 Min: 16 00:21:57.451 Number of Namespaces: 32 00:21:57.451 Compare Command: Supported 00:21:57.451 Write Uncorrectable Command: Not Supported 00:21:57.451 Dataset Management Command: Supported 00:21:57.451 Write Zeroes Command: Supported 00:21:57.451 Set Features Save Field: Not Supported 00:21:57.451 Reservations: Supported 00:21:57.451 Timestamp: Not Supported 00:21:57.451 Copy: Supported 00:21:57.451 Volatile Write Cache: Present 00:21:57.451 Atomic Write Unit (Normal): 1 00:21:57.451 Atomic Write Unit (PFail): 1 00:21:57.451 Atomic Compare & Write Unit: 1 00:21:57.451 Fused Compare & Write: Supported 00:21:57.451 Scatter-Gather List 00:21:57.451 SGL Command Set: Supported 00:21:57.451 SGL Keyed: Supported 00:21:57.451 SGL Bit Bucket Descriptor: Not Supported 00:21:57.451 SGL Metadata Pointer: Not Supported 00:21:57.451 Oversized SGL: Not Supported 00:21:57.451 SGL Metadata Address: Not Supported 00:21:57.451 SGL Offset: Supported 00:21:57.451 Transport SGL Data Block: Not Supported 00:21:57.451 Replay Protected Memory Block: Not Supported 00:21:57.451 00:21:57.451 Firmware Slot Information 00:21:57.451 ========================= 00:21:57.451 Active slot: 1 00:21:57.451 Slot 1 Firmware Revision: 24.09 00:21:57.451 00:21:57.451 00:21:57.451 Commands Supported and Effects 00:21:57.451 ============================== 00:21:57.451 Admin Commands 00:21:57.451 -------------- 00:21:57.451 Get Log Page (02h): Supported 00:21:57.451 Identify (06h): Supported 00:21:57.451 Abort (08h): Supported 00:21:57.451 Set Features (09h): Supported 00:21:57.451 Get Features (0Ah): Supported 00:21:57.451 Asynchronous Event Request (0Ch): Supported 00:21:57.451 Keep Alive (18h): Supported 00:21:57.451 I/O Commands 00:21:57.451 ------------ 00:21:57.451 Flush (00h): Supported LBA-Change 00:21:57.451 Write (01h): Supported LBA-Change 00:21:57.451 Read (02h): Supported 00:21:57.451 Compare (05h): Supported 00:21:57.451 Write Zeroes (08h): Supported LBA-Change 00:21:57.451 Dataset Management (09h): Supported LBA-Change 00:21:57.451 Copy (19h): Supported LBA-Change 00:21:57.451 00:21:57.451 Error Log 00:21:57.451 ========= 00:21:57.451 00:21:57.451 Arbitration 00:21:57.451 =========== 00:21:57.451 Arbitration Burst: 1 00:21:57.451 00:21:57.451 Power Management 00:21:57.451 ================ 00:21:57.451 Number of Power States: 1 00:21:57.451 Current Power State: Power State #0 00:21:57.451 Power State #0: 00:21:57.451 Max Power: 0.00 W 00:21:57.451 Non-Operational State: Operational 00:21:57.451 Entry Latency: Not Reported 00:21:57.451 Exit Latency: Not Reported 00:21:57.451 Relative Read Throughput: 0 00:21:57.451 Relative Read Latency: 0 00:21:57.451 Relative Write Throughput: 0 00:21:57.451 Relative Write Latency: 0 00:21:57.451 Idle Power: Not Reported 00:21:57.451 Active Power: Not Reported 00:21:57.451 Non-Operational Permissive Mode: Not Supported 00:21:57.451 00:21:57.451 Health Information 00:21:57.451 ================== 00:21:57.451 Critical Warnings: 00:21:57.451 Available Spare Space: OK 00:21:57.451 Temperature: OK 00:21:57.451 Device Reliability: OK 00:21:57.451 Read Only: No 00:21:57.451 Volatile Memory Backup: OK 00:21:57.451 Current Temperature: 0 Kelvin (-273 Celsius) 00:21:57.451 Temperature Threshold: 0 Kelvin (-273 Celsius) 00:21:57.451 Available Spare: 0% 00:21:57.451 Available Spare Threshold: 0% 00:21:57.451 Life Percentage Used:[2024-07-15 18:07:51.125526] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.125531] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=7 on tqpair(0x7a9ec0) 00:21:57.452 [2024-07-15 18:07:51.125537] nvme_qpair.c: 213:nvme_admin_qpair_print_command: *NOTICE*: GET FEATURES ERROR_RECOVERY cid:7 cdw10:00000005 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.452 [2024-07-15 18:07:51.125551] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d8c0, cid 7, qid 0 00:21:57.452 [2024-07-15 18:07:51.125697] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.452 [2024-07-15 18:07:51.125703] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.452 [2024-07-15 18:07:51.125706] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.125709] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d8c0) on tqpair=0x7a9ec0 00:21:57.452 [2024-07-15 18:07:51.125738] nvme_ctrlr.c:4357:nvme_ctrlr_destruct_async: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] Prepare to destruct SSD 00:21:57.452 [2024-07-15 18:07:51.125747] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82ce40) on tqpair=0x7a9ec0 00:21:57.452 [2024-07-15 18:07:51.125752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:57.452 [2024-07-15 18:07:51.125757] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82cfc0) on tqpair=0x7a9ec0 00:21:57.452 [2024-07-15 18:07:51.125761] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:57.452 [2024-07-15 18:07:51.125765] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d140) on tqpair=0x7a9ec0 00:21:57.452 [2024-07-15 18:07:51.125769] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:57.452 [2024-07-15 18:07:51.125773] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.452 [2024-07-15 18:07:51.125777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:21:57.452 [2024-07-15 18:07:51.125784] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.125787] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.125790] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.452 [2024-07-15 18:07:51.125796] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.452 [2024-07-15 18:07:51.125807] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.452 [2024-07-15 18:07:51.125879] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.452 [2024-07-15 18:07:51.125885] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.452 [2024-07-15 18:07:51.125890] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.125893] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.452 [2024-07-15 18:07:51.125899] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.125903] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.125905] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.452 [2024-07-15 18:07:51.125911] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY SET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.452 [2024-07-15 18:07:51.125924] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.452 [2024-07-15 18:07:51.126011] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.452 [2024-07-15 18:07:51.126016] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.452 [2024-07-15 18:07:51.126019] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126022] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.452 [2024-07-15 18:07:51.126026] nvme_ctrlr.c:1147:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] RTD3E = 0 us 00:21:57.452 [2024-07-15 18:07:51.126030] nvme_ctrlr.c:1150:nvme_ctrlr_shutdown_set_cc_done: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] shutdown timeout = 10000 ms 00:21:57.452 [2024-07-15 18:07:51.126038] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126042] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126045] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.452 [2024-07-15 18:07:51.126050] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.452 [2024-07-15 18:07:51.126059] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.452 [2024-07-15 18:07:51.126135] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.452 [2024-07-15 18:07:51.126140] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.452 [2024-07-15 18:07:51.126143] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126146] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.452 [2024-07-15 18:07:51.126154] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126158] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126161] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.452 [2024-07-15 18:07:51.126167] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.452 [2024-07-15 18:07:51.126175] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.452 [2024-07-15 18:07:51.126254] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.452 [2024-07-15 18:07:51.126260] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.452 [2024-07-15 18:07:51.126263] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126266] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.452 [2024-07-15 18:07:51.126274] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126277] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126280] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.452 [2024-07-15 18:07:51.126286] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.452 [2024-07-15 18:07:51.126295] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.452 [2024-07-15 18:07:51.126372] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.452 [2024-07-15 18:07:51.126378] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.452 [2024-07-15 18:07:51.126381] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126384] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.452 [2024-07-15 18:07:51.126391] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126395] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126398] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.452 [2024-07-15 18:07:51.126404] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.452 [2024-07-15 18:07:51.126413] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.452 [2024-07-15 18:07:51.126484] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.452 [2024-07-15 18:07:51.126489] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.452 [2024-07-15 18:07:51.126492] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126495] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.452 [2024-07-15 18:07:51.126503] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126507] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126510] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.452 [2024-07-15 18:07:51.126515] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.452 [2024-07-15 18:07:51.126524] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.452 [2024-07-15 18:07:51.126607] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.452 [2024-07-15 18:07:51.126613] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.452 [2024-07-15 18:07:51.126615] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126619] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.452 [2024-07-15 18:07:51.126627] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126630] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126633] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.452 [2024-07-15 18:07:51.126639] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.452 [2024-07-15 18:07:51.126648] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.452 [2024-07-15 18:07:51.126724] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.452 [2024-07-15 18:07:51.126730] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.452 [2024-07-15 18:07:51.126732] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126736] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.452 [2024-07-15 18:07:51.126744] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126747] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126750] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.452 [2024-07-15 18:07:51.126756] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.452 [2024-07-15 18:07:51.126765] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.452 [2024-07-15 18:07:51.126841] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.452 [2024-07-15 18:07:51.126846] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.452 [2024-07-15 18:07:51.126851] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126854] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.452 [2024-07-15 18:07:51.126862] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126865] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126868] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.452 [2024-07-15 18:07:51.126874] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.452 [2024-07-15 18:07:51.126883] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.452 [2024-07-15 18:07:51.126954] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.452 [2024-07-15 18:07:51.126959] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.452 [2024-07-15 18:07:51.126962] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126965] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.452 [2024-07-15 18:07:51.126973] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126977] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.452 [2024-07-15 18:07:51.126980] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.452 [2024-07-15 18:07:51.126985] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.453 [2024-07-15 18:07:51.126994] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.453 [2024-07-15 18:07:51.127075] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.453 [2024-07-15 18:07:51.127080] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.453 [2024-07-15 18:07:51.127083] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127087] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.453 [2024-07-15 18:07:51.127095] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127098] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127101] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.453 [2024-07-15 18:07:51.127107] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.453 [2024-07-15 18:07:51.127115] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.453 [2024-07-15 18:07:51.127192] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.453 [2024-07-15 18:07:51.127197] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.453 [2024-07-15 18:07:51.127200] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127203] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.453 [2024-07-15 18:07:51.127211] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127215] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127218] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.453 [2024-07-15 18:07:51.127223] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.453 [2024-07-15 18:07:51.127236] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.453 [2024-07-15 18:07:51.127308] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.453 [2024-07-15 18:07:51.127313] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.453 [2024-07-15 18:07:51.127316] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127321] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.453 [2024-07-15 18:07:51.127329] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127332] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127335] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.453 [2024-07-15 18:07:51.127341] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.453 [2024-07-15 18:07:51.127351] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.453 [2024-07-15 18:07:51.127422] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.453 [2024-07-15 18:07:51.127427] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.453 [2024-07-15 18:07:51.127430] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127433] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.453 [2024-07-15 18:07:51.127441] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127444] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127447] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.453 [2024-07-15 18:07:51.127453] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.453 [2024-07-15 18:07:51.127462] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.453 [2024-07-15 18:07:51.127543] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.453 [2024-07-15 18:07:51.127548] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.453 [2024-07-15 18:07:51.127551] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127554] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.453 [2024-07-15 18:07:51.127562] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127566] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127569] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.453 [2024-07-15 18:07:51.127574] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.453 [2024-07-15 18:07:51.127583] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.453 [2024-07-15 18:07:51.127658] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.453 [2024-07-15 18:07:51.127664] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.453 [2024-07-15 18:07:51.127667] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127670] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.453 [2024-07-15 18:07:51.127678] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127681] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127684] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.453 [2024-07-15 18:07:51.127690] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.453 [2024-07-15 18:07:51.127699] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.453 [2024-07-15 18:07:51.127775] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.453 [2024-07-15 18:07:51.127780] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.453 [2024-07-15 18:07:51.127783] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127787] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.453 [2024-07-15 18:07:51.127796] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127800] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127803] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.453 [2024-07-15 18:07:51.127808] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.453 [2024-07-15 18:07:51.127817] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.453 [2024-07-15 18:07:51.127888] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.453 [2024-07-15 18:07:51.127894] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.453 [2024-07-15 18:07:51.127897] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127900] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.453 [2024-07-15 18:07:51.127908] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127911] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.127914] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.453 [2024-07-15 18:07:51.127920] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.453 [2024-07-15 18:07:51.127929] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.453 [2024-07-15 18:07:51.128010] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.453 [2024-07-15 18:07:51.128015] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.453 [2024-07-15 18:07:51.128018] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.128021] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.453 [2024-07-15 18:07:51.128029] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.128033] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.128035] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.453 [2024-07-15 18:07:51.128041] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.453 [2024-07-15 18:07:51.128050] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.453 [2024-07-15 18:07:51.128126] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.453 [2024-07-15 18:07:51.128132] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.453 [2024-07-15 18:07:51.128135] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.128138] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.453 [2024-07-15 18:07:51.128146] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.128149] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.128152] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.453 [2024-07-15 18:07:51.128158] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.453 [2024-07-15 18:07:51.128167] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.453 [2024-07-15 18:07:51.128245] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.453 [2024-07-15 18:07:51.128251] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.453 [2024-07-15 18:07:51.128254] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.128257] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.453 [2024-07-15 18:07:51.128265] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.128270] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.128273] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.453 [2024-07-15 18:07:51.128279] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.453 [2024-07-15 18:07:51.128289] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.453 [2024-07-15 18:07:51.128360] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.453 [2024-07-15 18:07:51.128366] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.453 [2024-07-15 18:07:51.128369] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.128372] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.453 [2024-07-15 18:07:51.128380] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.128384] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.128387] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.453 [2024-07-15 18:07:51.128392] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.453 [2024-07-15 18:07:51.128401] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.453 [2024-07-15 18:07:51.128479] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.453 [2024-07-15 18:07:51.128485] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.453 [2024-07-15 18:07:51.128488] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.128491] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.453 [2024-07-15 18:07:51.128499] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.128502] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.453 [2024-07-15 18:07:51.128505] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.454 [2024-07-15 18:07:51.128511] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.454 [2024-07-15 18:07:51.128520] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.454 [2024-07-15 18:07:51.128596] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.454 [2024-07-15 18:07:51.128601] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.454 [2024-07-15 18:07:51.128604] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.128607] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.454 [2024-07-15 18:07:51.128615] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.128619] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.128622] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.454 [2024-07-15 18:07:51.128628] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.454 [2024-07-15 18:07:51.128637] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.454 [2024-07-15 18:07:51.128713] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.454 [2024-07-15 18:07:51.128718] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.454 [2024-07-15 18:07:51.128721] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.128725] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.454 [2024-07-15 18:07:51.128733] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.128736] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.128740] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.454 [2024-07-15 18:07:51.128746] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.454 [2024-07-15 18:07:51.128755] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.454 [2024-07-15 18:07:51.128829] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.454 [2024-07-15 18:07:51.128834] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.454 [2024-07-15 18:07:51.128837] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.128840] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.454 [2024-07-15 18:07:51.128848] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.128851] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.128855] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.454 [2024-07-15 18:07:51.128860] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.454 [2024-07-15 18:07:51.128869] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.454 [2024-07-15 18:07:51.128947] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.454 [2024-07-15 18:07:51.128952] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.454 [2024-07-15 18:07:51.128955] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.128958] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.454 [2024-07-15 18:07:51.128967] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.128970] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.128973] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.454 [2024-07-15 18:07:51.128979] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.454 [2024-07-15 18:07:51.128988] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.454 [2024-07-15 18:07:51.129063] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.454 [2024-07-15 18:07:51.129069] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.454 [2024-07-15 18:07:51.129072] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.129075] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.454 [2024-07-15 18:07:51.129083] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.129087] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.129090] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.454 [2024-07-15 18:07:51.129095] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.454 [2024-07-15 18:07:51.129104] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.454 [2024-07-15 18:07:51.129180] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.454 [2024-07-15 18:07:51.129185] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.454 [2024-07-15 18:07:51.129188] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.129191] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.454 [2024-07-15 18:07:51.129199] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.129203] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.129206] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.454 [2024-07-15 18:07:51.129213] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.454 [2024-07-15 18:07:51.129222] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.454 [2024-07-15 18:07:51.133238] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.454 [2024-07-15 18:07:51.133244] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.454 [2024-07-15 18:07:51.133247] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.133250] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.454 [2024-07-15 18:07:51.133261] nvme_tcp.c: 790:nvme_tcp_build_contig_request: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.133264] nvme_tcp.c: 967:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.133267] nvme_tcp.c: 976:nvme_tcp_qpair_capsule_cmd_send: *DEBUG*: capsule_cmd cid=3 on tqpair(0x7a9ec0) 00:21:57.454 [2024-07-15 18:07:51.133273] nvme_qpair.c: 218:nvme_admin_qpair_print_command: *NOTICE*: FABRIC PROPERTY GET qid:0 cid:3 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:21:57.454 [2024-07-15 18:07:51.133285] nvme_tcp.c: 941:nvme_tcp_qpair_cmd_send_complete: *DEBUG*: tcp req 0x82d2c0, cid 3, qid 0 00:21:57.454 [2024-07-15 18:07:51.133446] nvme_tcp.c:1187:nvme_tcp_pdu_ch_handle: *DEBUG*: pdu type = 5 00:21:57.454 [2024-07-15 18:07:51.133451] nvme_tcp.c:1975:nvme_tcp_pdu_psh_handle: *DEBUG*: enter: pdu type =5 00:21:57.454 [2024-07-15 18:07:51.133454] nvme_tcp.c:1648:nvme_tcp_capsule_resp_hdr_handle: *DEBUG*: enter 00:21:57.454 [2024-07-15 18:07:51.133457] nvme_tcp.c:1069:nvme_tcp_req_complete: *DEBUG*: complete tcp_req(0x82d2c0) on tqpair=0x7a9ec0 00:21:57.454 [2024-07-15 18:07:51.133464] nvme_ctrlr.c:1269:nvme_ctrlr_shutdown_poll_async: *DEBUG*: [nqn.2016-06.io.spdk:cnode1] shutdown complete in 7 milliseconds 00:21:57.454 0% 00:21:57.454 Data Units Read: 0 00:21:57.454 Data Units Written: 0 00:21:57.454 Host Read Commands: 0 00:21:57.454 Host Write Commands: 0 00:21:57.454 Controller Busy Time: 0 minutes 00:21:57.454 Power Cycles: 0 00:21:57.454 Power On Hours: 0 hours 00:21:57.454 Unsafe Shutdowns: 0 00:21:57.454 Unrecoverable Media Errors: 0 00:21:57.454 Lifetime Error Log Entries: 0 00:21:57.454 Warning Temperature Time: 0 minutes 00:21:57.454 Critical Temperature Time: 0 minutes 00:21:57.454 00:21:57.454 Number of Queues 00:21:57.454 ================ 00:21:57.454 Number of I/O Submission Queues: 127 00:21:57.454 Number of I/O Completion Queues: 127 00:21:57.454 00:21:57.454 Active Namespaces 00:21:57.454 ================= 00:21:57.454 Namespace ID:1 00:21:57.454 Error Recovery Timeout: Unlimited 00:21:57.454 Command Set Identifier: NVM (00h) 00:21:57.454 Deallocate: Supported 00:21:57.454 Deallocated/Unwritten Error: Not Supported 00:21:57.454 Deallocated Read Value: Unknown 00:21:57.454 Deallocate in Write Zeroes: Not Supported 00:21:57.454 Deallocated Guard Field: 0xFFFF 00:21:57.454 Flush: Supported 00:21:57.454 Reservation: Supported 00:21:57.454 Namespace Sharing Capabilities: Multiple Controllers 00:21:57.454 Size (in LBAs): 131072 (0GiB) 00:21:57.454 Capacity (in LBAs): 131072 (0GiB) 00:21:57.454 Utilization (in LBAs): 131072 (0GiB) 00:21:57.454 NGUID: ABCDEF0123456789ABCDEF0123456789 00:21:57.454 EUI64: ABCDEF0123456789 00:21:57.454 UUID: 98dd49eb-a0db-43a6-9107-626bc56df2fd 00:21:57.454 Thin Provisioning: Not Supported 00:21:57.454 Per-NS Atomic Units: Yes 00:21:57.454 Atomic Boundary Size (Normal): 0 00:21:57.454 Atomic Boundary Size (PFail): 0 00:21:57.454 Atomic Boundary Offset: 0 00:21:57.454 Maximum Single Source Range Length: 65535 00:21:57.454 Maximum Copy Length: 65535 00:21:57.454 Maximum Source Range Count: 1 00:21:57.454 NGUID/EUI64 Never Reused: No 00:21:57.454 Namespace Write Protected: No 00:21:57.454 Number of LBA Formats: 1 00:21:57.454 Current LBA Format: LBA Format #00 00:21:57.454 LBA Format #00: Data Size: 512 Metadata Size: 0 00:21:57.454 00:21:57.454 18:07:51 nvmf_tcp.nvmf_identify -- host/identify.sh@51 -- # sync 00:21:57.454 18:07:51 nvmf_tcp.nvmf_identify -- host/identify.sh@52 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:21:57.454 18:07:51 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@559 -- # xtrace_disable 00:21:57.454 18:07:51 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:21:57.454 18:07:51 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:21:57.454 18:07:51 nvmf_tcp.nvmf_identify -- host/identify.sh@54 -- # trap - SIGINT SIGTERM EXIT 00:21:57.454 18:07:51 nvmf_tcp.nvmf_identify -- host/identify.sh@56 -- # nvmftestfini 00:21:57.454 18:07:51 nvmf_tcp.nvmf_identify -- nvmf/common.sh@488 -- # nvmfcleanup 00:21:57.454 18:07:51 nvmf_tcp.nvmf_identify -- nvmf/common.sh@117 -- # sync 00:21:57.454 18:07:51 nvmf_tcp.nvmf_identify -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:21:57.454 18:07:51 nvmf_tcp.nvmf_identify -- nvmf/common.sh@120 -- # set +e 00:21:57.454 18:07:51 nvmf_tcp.nvmf_identify -- nvmf/common.sh@121 -- # for i in {1..20} 00:21:57.454 18:07:51 nvmf_tcp.nvmf_identify -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:21:57.454 rmmod nvme_tcp 00:21:57.733 rmmod nvme_fabrics 00:21:57.733 rmmod nvme_keyring 00:21:57.733 18:07:51 nvmf_tcp.nvmf_identify -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:21:57.733 18:07:51 nvmf_tcp.nvmf_identify -- nvmf/common.sh@124 -- # set -e 00:21:57.733 18:07:51 nvmf_tcp.nvmf_identify -- nvmf/common.sh@125 -- # return 0 00:21:57.733 18:07:51 nvmf_tcp.nvmf_identify -- nvmf/common.sh@489 -- # '[' -n 667870 ']' 00:21:57.733 18:07:51 nvmf_tcp.nvmf_identify -- nvmf/common.sh@490 -- # killprocess 667870 00:21:57.733 18:07:51 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@948 -- # '[' -z 667870 ']' 00:21:57.733 18:07:51 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@952 -- # kill -0 667870 00:21:57.733 18:07:51 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@953 -- # uname 00:21:57.733 18:07:51 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:21:57.733 18:07:51 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 667870 00:21:57.733 18:07:51 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:21:57.733 18:07:51 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:21:57.733 18:07:51 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@966 -- # echo 'killing process with pid 667870' 00:21:57.733 killing process with pid 667870 00:21:57.733 18:07:51 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@967 -- # kill 667870 00:21:57.733 18:07:51 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@972 -- # wait 667870 00:21:57.992 18:07:51 nvmf_tcp.nvmf_identify -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:21:57.992 18:07:51 nvmf_tcp.nvmf_identify -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:21:57.992 18:07:51 nvmf_tcp.nvmf_identify -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:21:57.992 18:07:51 nvmf_tcp.nvmf_identify -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:21:57.992 18:07:51 nvmf_tcp.nvmf_identify -- nvmf/common.sh@278 -- # remove_spdk_ns 00:21:57.992 18:07:51 nvmf_tcp.nvmf_identify -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:21:57.992 18:07:51 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:21:57.992 18:07:51 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:21:59.893 18:07:53 nvmf_tcp.nvmf_identify -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:21:59.893 00:21:59.893 real 0m9.337s 00:21:59.893 user 0m7.661s 00:21:59.893 sys 0m4.482s 00:21:59.893 18:07:53 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@1124 -- # xtrace_disable 00:21:59.893 18:07:53 nvmf_tcp.nvmf_identify -- common/autotest_common.sh@10 -- # set +x 00:21:59.893 ************************************ 00:21:59.893 END TEST nvmf_identify 00:21:59.893 ************************************ 00:21:59.893 18:07:53 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:21:59.893 18:07:53 nvmf_tcp -- nvmf/nvmf.sh@98 -- # run_test nvmf_perf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/perf.sh --transport=tcp 00:21:59.893 18:07:53 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:21:59.893 18:07:53 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:21:59.893 18:07:53 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:21:59.893 ************************************ 00:21:59.893 START TEST nvmf_perf 00:21:59.893 ************************************ 00:21:59.894 18:07:53 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/perf.sh --transport=tcp 00:22:00.173 * Looking for test storage... 00:22:00.173 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- host/perf.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@7 -- # uname -s 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- paths/export.sh@5 -- # export PATH 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@47 -- # : 0 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@51 -- # have_pci_nics=0 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- host/perf.sh@12 -- # MALLOC_BDEV_SIZE=64 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- host/perf.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- host/perf.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- host/perf.sh@17 -- # nvmftestinit 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@448 -- # prepare_net_devs 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@410 -- # local -g is_hw=no 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@412 -- # remove_spdk_ns 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- nvmf/common.sh@285 -- # xtrace_disable 00:22:00.173 18:07:53 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@291 -- # pci_devs=() 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@291 -- # local -a pci_devs 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@292 -- # pci_net_devs=() 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@293 -- # pci_drivers=() 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@293 -- # local -A pci_drivers 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@295 -- # net_devs=() 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@295 -- # local -ga net_devs 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@296 -- # e810=() 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@296 -- # local -ga e810 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@297 -- # x722=() 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@297 -- # local -ga x722 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@298 -- # mlx=() 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@298 -- # local -ga mlx 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:22:05.469 Found 0000:86:00.0 (0x8086 - 0x159b) 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:22:05.469 Found 0000:86:00.1 (0x8086 - 0x159b) 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@390 -- # [[ up == up ]] 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:22:05.469 Found net devices under 0000:86:00.0: cvl_0_0 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@390 -- # [[ up == up ]] 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:22:05.469 Found net devices under 0000:86:00.1: cvl_0_1 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:22:05.469 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@414 -- # is_hw=yes 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:22:05.470 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:22:05.470 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.185 ms 00:22:05.470 00:22:05.470 --- 10.0.0.2 ping statistics --- 00:22:05.470 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:05.470 rtt min/avg/max/mdev = 0.185/0.185/0.185/0.000 ms 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:22:05.470 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:22:05.470 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.202 ms 00:22:05.470 00:22:05.470 --- 10.0.0.1 ping statistics --- 00:22:05.470 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:05.470 rtt min/avg/max/mdev = 0.202/0.202/0.202/0.000 ms 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@422 -- # return 0 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- host/perf.sh@18 -- # nvmfappstart -m 0xF 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@722 -- # xtrace_disable 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@481 -- # nvmfpid=671505 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@482 -- # waitforlisten 671505 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@829 -- # '[' -z 671505 ']' 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@834 -- # local max_retries=100 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:05.470 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@838 -- # xtrace_disable 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:22:05.470 18:07:58 nvmf_tcp.nvmf_perf -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:22:05.470 [2024-07-15 18:07:58.759171] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:22:05.470 [2024-07-15 18:07:58.759215] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:22:05.470 EAL: No free 2048 kB hugepages reported on node 1 00:22:05.470 [2024-07-15 18:07:58.815497] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:22:05.470 [2024-07-15 18:07:58.896067] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:22:05.470 [2024-07-15 18:07:58.896104] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:22:05.470 [2024-07-15 18:07:58.896111] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:22:05.470 [2024-07-15 18:07:58.896117] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:22:05.470 [2024-07-15 18:07:58.896122] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:22:05.470 [2024-07-15 18:07:58.896161] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:22:05.470 [2024-07-15 18:07:58.896178] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:22:05.470 [2024-07-15 18:07:58.896267] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:22:05.470 [2024-07-15 18:07:58.896269] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:22:06.036 18:07:59 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:22:06.036 18:07:59 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@862 -- # return 0 00:22:06.036 18:07:59 nvmf_tcp.nvmf_perf -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:22:06.036 18:07:59 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@728 -- # xtrace_disable 00:22:06.036 18:07:59 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:22:06.036 18:07:59 nvmf_tcp.nvmf_perf -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:22:06.036 18:07:59 nvmf_tcp.nvmf_perf -- host/perf.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:22:06.036 18:07:59 nvmf_tcp.nvmf_perf -- host/perf.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py load_subsystem_config 00:22:09.319 18:08:02 nvmf_tcp.nvmf_perf -- host/perf.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py framework_get_config bdev 00:22:09.319 18:08:02 nvmf_tcp.nvmf_perf -- host/perf.sh@30 -- # jq -r '.[].params | select(.name=="Nvme0").traddr' 00:22:09.319 18:08:02 nvmf_tcp.nvmf_perf -- host/perf.sh@30 -- # local_nvme_trid=0000:5e:00.0 00:22:09.319 18:08:02 nvmf_tcp.nvmf_perf -- host/perf.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 00:22:09.319 18:08:03 nvmf_tcp.nvmf_perf -- host/perf.sh@31 -- # bdevs=' Malloc0' 00:22:09.319 18:08:03 nvmf_tcp.nvmf_perf -- host/perf.sh@33 -- # '[' -n 0000:5e:00.0 ']' 00:22:09.319 18:08:03 nvmf_tcp.nvmf_perf -- host/perf.sh@34 -- # bdevs=' Malloc0 Nvme0n1' 00:22:09.319 18:08:03 nvmf_tcp.nvmf_perf -- host/perf.sh@37 -- # '[' tcp == rdma ']' 00:22:09.319 18:08:03 nvmf_tcp.nvmf_perf -- host/perf.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o 00:22:09.577 [2024-07-15 18:08:03.163746] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:22:09.577 18:08:03 nvmf_tcp.nvmf_perf -- host/perf.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:22:09.836 18:08:03 nvmf_tcp.nvmf_perf -- host/perf.sh@45 -- # for bdev in $bdevs 00:22:09.836 18:08:03 nvmf_tcp.nvmf_perf -- host/perf.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:22:09.836 18:08:03 nvmf_tcp.nvmf_perf -- host/perf.sh@45 -- # for bdev in $bdevs 00:22:09.836 18:08:03 nvmf_tcp.nvmf_perf -- host/perf.sh@46 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Nvme0n1 00:22:10.093 18:08:03 nvmf_tcp.nvmf_perf -- host/perf.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:22:10.351 [2024-07-15 18:08:03.882424] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:22:10.351 18:08:03 nvmf_tcp.nvmf_perf -- host/perf.sh@49 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:22:10.627 18:08:04 nvmf_tcp.nvmf_perf -- host/perf.sh@52 -- # '[' -n 0000:5e:00.0 ']' 00:22:10.627 18:08:04 nvmf_tcp.nvmf_perf -- host/perf.sh@53 -- # perf_app -i 0 -q 32 -o 4096 -w randrw -M 50 -t 1 -r 'trtype:PCIe traddr:0000:5e:00.0' 00:22:10.627 18:08:04 nvmf_tcp.nvmf_perf -- host/perf.sh@21 -- # '[' 0 -eq 1 ']' 00:22:10.627 18:08:04 nvmf_tcp.nvmf_perf -- host/perf.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -i 0 -q 32 -o 4096 -w randrw -M 50 -t 1 -r 'trtype:PCIe traddr:0000:5e:00.0' 00:22:12.004 Initializing NVMe Controllers 00:22:12.004 Attached to NVMe Controller at 0000:5e:00.0 [8086:0a54] 00:22:12.004 Associating PCIE (0000:5e:00.0) NSID 1 with lcore 0 00:22:12.004 Initialization complete. Launching workers. 00:22:12.004 ======================================================== 00:22:12.004 Latency(us) 00:22:12.004 Device Information : IOPS MiB/s Average min max 00:22:12.004 PCIE (0000:5e:00.0) NSID 1 from core 0: 97865.04 382.29 326.42 39.50 4468.79 00:22:12.004 ======================================================== 00:22:12.004 Total : 97865.04 382.29 326.42 39.50 4468.79 00:22:12.004 00:22:12.004 18:08:05 nvmf_tcp.nvmf_perf -- host/perf.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 1 -o 4096 -w randrw -M 50 -t 1 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:22:12.004 EAL: No free 2048 kB hugepages reported on node 1 00:22:12.939 Initializing NVMe Controllers 00:22:12.939 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:22:12.939 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:22:12.939 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:22:12.939 Initialization complete. Launching workers. 00:22:12.939 ======================================================== 00:22:12.939 Latency(us) 00:22:12.939 Device Information : IOPS MiB/s Average min max 00:22:12.939 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 143.00 0.56 7158.79 141.71 45655.01 00:22:12.939 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 51.00 0.20 20497.91 7946.53 47918.50 00:22:12.939 ======================================================== 00:22:12.939 Total : 194.00 0.76 10665.47 141.71 47918.50 00:22:12.939 00:22:12.939 18:08:06 nvmf_tcp.nvmf_perf -- host/perf.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 32 -o 4096 -w randrw -M 50 -t 1 -HI -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:22:12.939 EAL: No free 2048 kB hugepages reported on node 1 00:22:14.315 Initializing NVMe Controllers 00:22:14.315 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:22:14.315 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:22:14.315 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:22:14.315 Initialization complete. Launching workers. 00:22:14.315 ======================================================== 00:22:14.315 Latency(us) 00:22:14.315 Device Information : IOPS MiB/s Average min max 00:22:14.315 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 10857.99 42.41 2949.07 357.99 6305.40 00:22:14.315 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 3914.00 15.29 8219.47 7115.70 15837.52 00:22:14.315 ======================================================== 00:22:14.315 Total : 14771.99 57.70 4345.52 357.99 15837.52 00:22:14.315 00:22:14.315 18:08:07 nvmf_tcp.nvmf_perf -- host/perf.sh@59 -- # [[ e810 == \e\8\1\0 ]] 00:22:14.315 18:08:07 nvmf_tcp.nvmf_perf -- host/perf.sh@59 -- # [[ tcp == \r\d\m\a ]] 00:22:14.315 18:08:07 nvmf_tcp.nvmf_perf -- host/perf.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 262144 -O 16384 -w randrw -M 50 -t 2 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:22:14.315 EAL: No free 2048 kB hugepages reported on node 1 00:22:16.847 Initializing NVMe Controllers 00:22:16.847 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:22:16.847 Controller IO queue size 128, less than required. 00:22:16.847 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:22:16.847 Controller IO queue size 128, less than required. 00:22:16.847 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:22:16.847 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:22:16.847 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:22:16.847 Initialization complete. Launching workers. 00:22:16.847 ======================================================== 00:22:16.847 Latency(us) 00:22:16.847 Device Information : IOPS MiB/s Average min max 00:22:16.847 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 1248.11 312.03 104967.34 56592.20 163877.37 00:22:16.847 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 585.85 146.46 225894.10 69419.06 351579.76 00:22:16.847 ======================================================== 00:22:16.847 Total : 1833.96 458.49 143596.72 56592.20 351579.76 00:22:16.847 00:22:16.847 18:08:10 nvmf_tcp.nvmf_perf -- host/perf.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 36964 -O 4096 -w randrw -M 50 -t 5 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' -c 0xf -P 4 00:22:16.847 EAL: No free 2048 kB hugepages reported on node 1 00:22:17.104 No valid NVMe controllers or AIO or URING devices found 00:22:17.104 Initializing NVMe Controllers 00:22:17.104 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:22:17.104 Controller IO queue size 128, less than required. 00:22:17.104 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:22:17.104 WARNING: IO size 36964 (-o) is not a multiple of nsid 1 sector size 512. Removing this ns from test 00:22:17.104 Controller IO queue size 128, less than required. 00:22:17.104 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:22:17.104 WARNING: IO size 36964 (-o) is not a multiple of nsid 2 sector size 512. Removing this ns from test 00:22:17.104 WARNING: Some requested NVMe devices were skipped 00:22:17.104 18:08:10 nvmf_tcp.nvmf_perf -- host/perf.sh@65 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_perf -q 128 -o 262144 -w randrw -M 50 -t 2 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' --transport-stat 00:22:17.104 EAL: No free 2048 kB hugepages reported on node 1 00:22:19.638 Initializing NVMe Controllers 00:22:19.638 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:22:19.638 Controller IO queue size 128, less than required. 00:22:19.638 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:22:19.638 Controller IO queue size 128, less than required. 00:22:19.638 Consider using lower queue depth or smaller IO size, because IO requests may be queued at the NVMe driver. 00:22:19.638 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 with lcore 0 00:22:19.638 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 with lcore 0 00:22:19.638 Initialization complete. Launching workers. 00:22:19.638 00:22:19.638 ==================== 00:22:19.638 lcore 0, ns TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 statistics: 00:22:19.638 TCP transport: 00:22:19.638 polls: 34331 00:22:19.638 idle_polls: 15186 00:22:19.638 sock_completions: 19145 00:22:19.638 nvme_completions: 5109 00:22:19.638 submitted_requests: 7660 00:22:19.638 queued_requests: 1 00:22:19.638 00:22:19.638 ==================== 00:22:19.638 lcore 0, ns TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 statistics: 00:22:19.638 TCP transport: 00:22:19.638 polls: 33920 00:22:19.638 idle_polls: 13969 00:22:19.638 sock_completions: 19951 00:22:19.638 nvme_completions: 5225 00:22:19.638 submitted_requests: 7772 00:22:19.638 queued_requests: 1 00:22:19.638 ======================================================== 00:22:19.638 Latency(us) 00:22:19.638 Device Information : IOPS MiB/s Average min max 00:22:19.638 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 1 from core 0: 1276.97 319.24 103019.29 61748.91 171295.98 00:22:19.638 TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) NSID 2 from core 0: 1305.97 326.49 99339.32 34962.81 133182.26 00:22:19.638 ======================================================== 00:22:19.638 Total : 2582.94 645.74 101158.65 34962.81 171295.98 00:22:19.638 00:22:19.638 18:08:13 nvmf_tcp.nvmf_perf -- host/perf.sh@66 -- # sync 00:22:19.638 18:08:13 nvmf_tcp.nvmf_perf -- host/perf.sh@67 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:22:19.897 18:08:13 nvmf_tcp.nvmf_perf -- host/perf.sh@69 -- # '[' 0 -eq 1 ']' 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- host/perf.sh@112 -- # trap - SIGINT SIGTERM EXIT 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- host/perf.sh@114 -- # nvmftestfini 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- nvmf/common.sh@488 -- # nvmfcleanup 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- nvmf/common.sh@117 -- # sync 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- nvmf/common.sh@120 -- # set +e 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- nvmf/common.sh@121 -- # for i in {1..20} 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:22:19.898 rmmod nvme_tcp 00:22:19.898 rmmod nvme_fabrics 00:22:19.898 rmmod nvme_keyring 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- nvmf/common.sh@124 -- # set -e 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- nvmf/common.sh@125 -- # return 0 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- nvmf/common.sh@489 -- # '[' -n 671505 ']' 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- nvmf/common.sh@490 -- # killprocess 671505 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@948 -- # '[' -z 671505 ']' 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@952 -- # kill -0 671505 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@953 -- # uname 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 671505 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@966 -- # echo 'killing process with pid 671505' 00:22:19.898 killing process with pid 671505 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@967 -- # kill 671505 00:22:19.898 18:08:13 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@972 -- # wait 671505 00:22:21.801 18:08:15 nvmf_tcp.nvmf_perf -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:22:21.801 18:08:15 nvmf_tcp.nvmf_perf -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:22:21.801 18:08:15 nvmf_tcp.nvmf_perf -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:22:21.801 18:08:15 nvmf_tcp.nvmf_perf -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:22:21.801 18:08:15 nvmf_tcp.nvmf_perf -- nvmf/common.sh@278 -- # remove_spdk_ns 00:22:21.801 18:08:15 nvmf_tcp.nvmf_perf -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:21.801 18:08:15 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:22:21.801 18:08:15 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:23.704 18:08:17 nvmf_tcp.nvmf_perf -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:22:23.704 00:22:23.704 real 0m23.504s 00:22:23.704 user 1m4.358s 00:22:23.704 sys 0m6.793s 00:22:23.704 18:08:17 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@1124 -- # xtrace_disable 00:22:23.704 18:08:17 nvmf_tcp.nvmf_perf -- common/autotest_common.sh@10 -- # set +x 00:22:23.704 ************************************ 00:22:23.704 END TEST nvmf_perf 00:22:23.704 ************************************ 00:22:23.704 18:08:17 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:22:23.704 18:08:17 nvmf_tcp -- nvmf/nvmf.sh@99 -- # run_test nvmf_fio_host /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/fio.sh --transport=tcp 00:22:23.704 18:08:17 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:22:23.704 18:08:17 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:22:23.704 18:08:17 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:22:23.704 ************************************ 00:22:23.704 START TEST nvmf_fio_host 00:22:23.704 ************************************ 00:22:23.704 18:08:17 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/fio.sh --transport=tcp 00:22:23.704 * Looking for test storage... 00:22:23.704 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:22:23.704 18:08:17 nvmf_tcp.nvmf_fio_host -- host/fio.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:22:23.704 18:08:17 nvmf_tcp.nvmf_fio_host -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:22:23.704 18:08:17 nvmf_tcp.nvmf_fio_host -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:22:23.704 18:08:17 nvmf_tcp.nvmf_fio_host -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:22:23.704 18:08:17 nvmf_tcp.nvmf_fio_host -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:23.704 18:08:17 nvmf_tcp.nvmf_fio_host -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:23.704 18:08:17 nvmf_tcp.nvmf_fio_host -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:23.704 18:08:17 nvmf_tcp.nvmf_fio_host -- paths/export.sh@5 -- # export PATH 00:22:23.704 18:08:17 nvmf_tcp.nvmf_fio_host -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:23.704 18:08:17 nvmf_tcp.nvmf_fio_host -- host/fio.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:22:23.704 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@7 -- # uname -s 00:22:23.704 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- paths/export.sh@5 -- # export PATH 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@47 -- # : 0 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@51 -- # have_pci_nics=0 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- host/fio.sh@12 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- host/fio.sh@14 -- # nvmftestinit 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@448 -- # prepare_net_devs 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@410 -- # local -g is_hw=no 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@412 -- # remove_spdk_ns 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@285 -- # xtrace_disable 00:22:23.705 18:08:17 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@291 -- # pci_devs=() 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@291 -- # local -a pci_devs 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@292 -- # pci_net_devs=() 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@293 -- # pci_drivers=() 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@293 -- # local -A pci_drivers 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@295 -- # net_devs=() 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@295 -- # local -ga net_devs 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@296 -- # e810=() 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@296 -- # local -ga e810 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@297 -- # x722=() 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@297 -- # local -ga x722 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@298 -- # mlx=() 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@298 -- # local -ga mlx 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:22:29.023 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:22:29.024 Found 0000:86:00.0 (0x8086 - 0x159b) 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:22:29.024 Found 0000:86:00.1 (0x8086 - 0x159b) 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@390 -- # [[ up == up ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:22:29.024 Found net devices under 0000:86:00.0: cvl_0_0 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@390 -- # [[ up == up ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:22:29.024 Found net devices under 0000:86:00.1: cvl_0_1 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@414 -- # is_hw=yes 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:22:29.024 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:22:29.024 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.201 ms 00:22:29.024 00:22:29.024 --- 10.0.0.2 ping statistics --- 00:22:29.024 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:29.024 rtt min/avg/max/mdev = 0.201/0.201/0.201/0.000 ms 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:22:29.024 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:22:29.024 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.184 ms 00:22:29.024 00:22:29.024 --- 10.0.0.1 ping statistics --- 00:22:29.024 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:29.024 rtt min/avg/max/mdev = 0.184/0.184/0.184/0.000 ms 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@422 -- # return 0 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- host/fio.sh@16 -- # [[ y != y ]] 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- host/fio.sh@21 -- # timing_enter start_nvmf_tgt 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@722 -- # xtrace_disable 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- host/fio.sh@24 -- # nvmfpid=677610 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- host/fio.sh@23 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- host/fio.sh@26 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- host/fio.sh@28 -- # waitforlisten 677610 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@829 -- # '[' -z 677610 ']' 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@834 -- # local max_retries=100 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:29.024 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@838 -- # xtrace_disable 00:22:29.024 18:08:22 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:22:29.283 [2024-07-15 18:08:22.790943] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:22:29.283 [2024-07-15 18:08:22.790991] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:22:29.283 EAL: No free 2048 kB hugepages reported on node 1 00:22:29.283 [2024-07-15 18:08:22.848616] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:22:29.283 [2024-07-15 18:08:22.929876] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:22:29.283 [2024-07-15 18:08:22.929911] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:22:29.283 [2024-07-15 18:08:22.929918] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:22:29.283 [2024-07-15 18:08:22.929924] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:22:29.283 [2024-07-15 18:08:22.929929] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:22:29.283 [2024-07-15 18:08:22.929971] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:22:29.283 [2024-07-15 18:08:22.930068] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:22:29.283 [2024-07-15 18:08:22.930153] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:22:29.283 [2024-07-15 18:08:22.930154] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:22:30.217 18:08:23 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:22:30.217 18:08:23 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@862 -- # return 0 00:22:30.217 18:08:23 nvmf_tcp.nvmf_fio_host -- host/fio.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:22:30.217 [2024-07-15 18:08:23.750484] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:22:30.217 18:08:23 nvmf_tcp.nvmf_fio_host -- host/fio.sh@30 -- # timing_exit start_nvmf_tgt 00:22:30.217 18:08:23 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@728 -- # xtrace_disable 00:22:30.217 18:08:23 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:22:30.217 18:08:23 nvmf_tcp.nvmf_fio_host -- host/fio.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc1 00:22:30.475 Malloc1 00:22:30.475 18:08:24 nvmf_tcp.nvmf_fio_host -- host/fio.sh@33 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:22:30.475 18:08:24 nvmf_tcp.nvmf_fio_host -- host/fio.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc1 00:22:30.732 18:08:24 nvmf_tcp.nvmf_fio_host -- host/fio.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:22:30.989 [2024-07-15 18:08:24.532596] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:22:30.989 18:08:24 nvmf_tcp.nvmf_fio_host -- host/fio.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- host/fio.sh@38 -- # PLUGIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- host/fio.sh@41 -- # fio_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1360 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1339 -- # local sanitizers 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1341 -- # shift 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1343 -- # local asan_lib= 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1345 -- # grep libasan 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1345 -- # asan_lib= 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1345 -- # asan_lib= 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme' 00:22:31.262 18:08:24 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/example_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' --bs=4096 00:22:31.527 test: (g=0): rw=randrw, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk, iodepth=128 00:22:31.527 fio-3.35 00:22:31.527 Starting 1 thread 00:22:31.527 EAL: No free 2048 kB hugepages reported on node 1 00:22:34.056 00:22:34.056 test: (groupid=0, jobs=1): err= 0: pid=678112: Mon Jul 15 18:08:27 2024 00:22:34.056 read: IOPS=11.8k, BW=46.3MiB/s (48.5MB/s)(92.8MiB/2005msec) 00:22:34.056 slat (nsec): min=1601, max=241516, avg=1750.66, stdev=2270.11 00:22:34.056 clat (usec): min=3105, max=10309, avg=5978.14, stdev=465.26 00:22:34.056 lat (usec): min=3139, max=10311, avg=5979.89, stdev=465.23 00:22:34.056 clat percentiles (usec): 00:22:34.056 | 1.00th=[ 4817], 5.00th=[ 5276], 10.00th=[ 5407], 20.00th=[ 5604], 00:22:34.056 | 30.00th=[ 5735], 40.00th=[ 5866], 50.00th=[ 5997], 60.00th=[ 6128], 00:22:34.056 | 70.00th=[ 6194], 80.00th=[ 6325], 90.00th=[ 6521], 95.00th=[ 6652], 00:22:34.056 | 99.00th=[ 6980], 99.50th=[ 7177], 99.90th=[ 8455], 99.95th=[ 9634], 00:22:34.056 | 99.99th=[10290] 00:22:34.056 bw ( KiB/s): min=46624, max=47816, per=99.95%, avg=47378.00, stdev=562.64, samples=4 00:22:34.056 iops : min=11656, max=11954, avg=11844.50, stdev=140.66, samples=4 00:22:34.056 write: IOPS=11.8k, BW=46.1MiB/s (48.3MB/s)(92.4MiB/2005msec); 0 zone resets 00:22:34.056 slat (nsec): min=1641, max=227703, avg=1839.90, stdev=1664.08 00:22:34.056 clat (usec): min=2463, max=9540, avg=4818.91, stdev=383.72 00:22:34.056 lat (usec): min=2479, max=9542, avg=4820.75, stdev=383.76 00:22:34.056 clat percentiles (usec): 00:22:34.056 | 1.00th=[ 3949], 5.00th=[ 4228], 10.00th=[ 4359], 20.00th=[ 4555], 00:22:34.056 | 30.00th=[ 4621], 40.00th=[ 4752], 50.00th=[ 4817], 60.00th=[ 4883], 00:22:34.056 | 70.00th=[ 5014], 80.00th=[ 5080], 90.00th=[ 5276], 95.00th=[ 5407], 00:22:34.056 | 99.00th=[ 5669], 99.50th=[ 5866], 99.90th=[ 7570], 99.95th=[ 8356], 00:22:34.056 | 99.99th=[ 9503] 00:22:34.056 bw ( KiB/s): min=46720, max=47616, per=100.00%, avg=47184.00, stdev=366.31, samples=4 00:22:34.056 iops : min=11680, max=11904, avg=11796.00, stdev=91.58, samples=4 00:22:34.056 lat (msec) : 4=0.72%, 10=99.26%, 20=0.01% 00:22:34.056 cpu : usr=69.56%, sys=27.00%, ctx=77, majf=0, minf=6 00:22:34.056 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.1%, 16=0.1%, 32=0.1%, >=64=99.9% 00:22:34.056 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:22:34.056 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:22:34.056 issued rwts: total=23761,23647,0,0 short=0,0,0,0 dropped=0,0,0,0 00:22:34.056 latency : target=0, window=0, percentile=100.00%, depth=128 00:22:34.056 00:22:34.056 Run status group 0 (all jobs): 00:22:34.056 READ: bw=46.3MiB/s (48.5MB/s), 46.3MiB/s-46.3MiB/s (48.5MB/s-48.5MB/s), io=92.8MiB (97.3MB), run=2005-2005msec 00:22:34.056 WRITE: bw=46.1MiB/s (48.3MB/s), 46.1MiB/s-46.1MiB/s (48.3MB/s-48.3MB/s), io=92.4MiB (96.9MB), run=2005-2005msec 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- host/fio.sh@45 -- # fio_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/mock_sgl_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1360 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/mock_sgl_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1339 -- # local sanitizers 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1341 -- # shift 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1343 -- # local asan_lib= 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1345 -- # grep libasan 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1345 -- # asan_lib= 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1345 -- # asan_lib= 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_nvme' 00:22:34.057 18:08:27 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme/mock_sgl_config.fio '--filename=trtype=tcp adrfam=IPv4 traddr=10.0.0.2 trsvcid=4420 ns=1' 00:22:34.057 test: (g=0): rw=randrw, bs=(R) 16.0KiB-16.0KiB, (W) 16.0KiB-16.0KiB, (T) 16.0KiB-16.0KiB, ioengine=spdk, iodepth=128 00:22:34.057 fio-3.35 00:22:34.057 Starting 1 thread 00:22:34.315 EAL: No free 2048 kB hugepages reported on node 1 00:22:36.844 00:22:36.844 test: (groupid=0, jobs=1): err= 0: pid=678683: Mon Jul 15 18:08:30 2024 00:22:36.844 read: IOPS=10.8k, BW=168MiB/s (176MB/s)(337MiB/2006msec) 00:22:36.844 slat (usec): min=2, max=106, avg= 2.92, stdev= 1.68 00:22:36.844 clat (usec): min=2462, max=13359, avg=7056.01, stdev=1693.13 00:22:36.844 lat (usec): min=2465, max=13362, avg=7058.94, stdev=1693.26 00:22:36.844 clat percentiles (usec): 00:22:36.844 | 1.00th=[ 3752], 5.00th=[ 4424], 10.00th=[ 4817], 20.00th=[ 5538], 00:22:36.844 | 30.00th=[ 6063], 40.00th=[ 6521], 50.00th=[ 7046], 60.00th=[ 7635], 00:22:36.844 | 70.00th=[ 8029], 80.00th=[ 8356], 90.00th=[ 9110], 95.00th=[ 9765], 00:22:36.844 | 99.00th=[11600], 99.50th=[12256], 99.90th=[13304], 99.95th=[13304], 00:22:36.844 | 99.99th=[13304] 00:22:36.844 bw ( KiB/s): min=77632, max=94080, per=50.05%, avg=86160.00, stdev=6766.59, samples=4 00:22:36.844 iops : min= 4852, max= 5880, avg=5385.00, stdev=422.91, samples=4 00:22:36.844 write: IOPS=6391, BW=99.9MiB/s (105MB/s)(176MiB/1765msec); 0 zone resets 00:22:36.844 slat (usec): min=30, max=379, avg=32.54, stdev= 8.04 00:22:36.844 clat (usec): min=2578, max=16064, avg=8545.94, stdev=1456.62 00:22:36.844 lat (usec): min=2608, max=16095, avg=8578.47, stdev=1458.12 00:22:36.844 clat percentiles (usec): 00:22:36.844 | 1.00th=[ 5866], 5.00th=[ 6587], 10.00th=[ 6915], 20.00th=[ 7373], 00:22:36.844 | 30.00th=[ 7701], 40.00th=[ 8029], 50.00th=[ 8291], 60.00th=[ 8717], 00:22:36.844 | 70.00th=[ 8979], 80.00th=[ 9634], 90.00th=[10552], 95.00th=[11338], 00:22:36.844 | 99.00th=[12649], 99.50th=[13173], 99.90th=[14353], 99.95th=[14615], 00:22:36.844 | 99.99th=[16057] 00:22:36.844 bw ( KiB/s): min=80352, max=98304, per=87.73%, avg=89712.00, stdev=7379.94, samples=4 00:22:36.844 iops : min= 5022, max= 6144, avg=5607.00, stdev=461.25, samples=4 00:22:36.844 lat (msec) : 4=1.39%, 10=90.63%, 20=7.98% 00:22:36.844 cpu : usr=85.19%, sys=13.02%, ctx=36, majf=0, minf=3 00:22:36.844 IO depths : 1=0.1%, 2=0.1%, 4=0.1%, 8=0.2%, 16=0.3%, 32=0.7%, >=64=98.7% 00:22:36.844 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:22:36.844 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1% 00:22:36.844 issued rwts: total=21583,11281,0,0 short=0,0,0,0 dropped=0,0,0,0 00:22:36.844 latency : target=0, window=0, percentile=100.00%, depth=128 00:22:36.844 00:22:36.844 Run status group 0 (all jobs): 00:22:36.844 READ: bw=168MiB/s (176MB/s), 168MiB/s-168MiB/s (176MB/s-176MB/s), io=337MiB (354MB), run=2006-2006msec 00:22:36.844 WRITE: bw=99.9MiB/s (105MB/s), 99.9MiB/s-99.9MiB/s (105MB/s-105MB/s), io=176MiB (185MB), run=1765-1765msec 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- host/fio.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- host/fio.sh@49 -- # '[' 0 -eq 1 ']' 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- host/fio.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- host/fio.sh@85 -- # rm -f ./local-test-0-verify.state 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- host/fio.sh@86 -- # nvmftestfini 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@488 -- # nvmfcleanup 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@117 -- # sync 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@120 -- # set +e 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@121 -- # for i in {1..20} 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:22:36.844 rmmod nvme_tcp 00:22:36.844 rmmod nvme_fabrics 00:22:36.844 rmmod nvme_keyring 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@124 -- # set -e 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@125 -- # return 0 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@489 -- # '[' -n 677610 ']' 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@490 -- # killprocess 677610 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@948 -- # '[' -z 677610 ']' 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@952 -- # kill -0 677610 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@953 -- # uname 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 677610 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@966 -- # echo 'killing process with pid 677610' 00:22:36.844 killing process with pid 677610 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@967 -- # kill 677610 00:22:36.844 18:08:30 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@972 -- # wait 677610 00:22:37.103 18:08:30 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:22:37.103 18:08:30 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:22:37.103 18:08:30 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:22:37.103 18:08:30 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:22:37.103 18:08:30 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@278 -- # remove_spdk_ns 00:22:37.103 18:08:30 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:37.103 18:08:30 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:22:37.103 18:08:30 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:39.003 18:08:32 nvmf_tcp.nvmf_fio_host -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:22:39.003 00:22:39.003 real 0m15.532s 00:22:39.003 user 0m47.519s 00:22:39.003 sys 0m6.080s 00:22:39.003 18:08:32 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@1124 -- # xtrace_disable 00:22:39.003 18:08:32 nvmf_tcp.nvmf_fio_host -- common/autotest_common.sh@10 -- # set +x 00:22:39.003 ************************************ 00:22:39.003 END TEST nvmf_fio_host 00:22:39.003 ************************************ 00:22:39.261 18:08:32 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:22:39.261 18:08:32 nvmf_tcp -- nvmf/nvmf.sh@100 -- # run_test nvmf_failover /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/failover.sh --transport=tcp 00:22:39.261 18:08:32 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:22:39.261 18:08:32 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:22:39.261 18:08:32 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:22:39.261 ************************************ 00:22:39.261 START TEST nvmf_failover 00:22:39.261 ************************************ 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/failover.sh --transport=tcp 00:22:39.261 * Looking for test storage... 00:22:39.261 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- host/failover.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@7 -- # uname -s 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- paths/export.sh@5 -- # export PATH 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@47 -- # : 0 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@51 -- # have_pci_nics=0 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- host/failover.sh@11 -- # MALLOC_BDEV_SIZE=64 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- host/failover.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- host/failover.sh@14 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- host/failover.sh@16 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- host/failover.sh@18 -- # nvmftestinit 00:22:39.261 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:22:39.262 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:22:39.262 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@448 -- # prepare_net_devs 00:22:39.262 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@410 -- # local -g is_hw=no 00:22:39.262 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@412 -- # remove_spdk_ns 00:22:39.262 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:22:39.262 18:08:32 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:22:39.262 18:08:32 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:22:39.262 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:22:39.262 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:22:39.262 18:08:32 nvmf_tcp.nvmf_failover -- nvmf/common.sh@285 -- # xtrace_disable 00:22:39.262 18:08:32 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@291 -- # pci_devs=() 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@291 -- # local -a pci_devs 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@292 -- # pci_net_devs=() 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@293 -- # pci_drivers=() 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@293 -- # local -A pci_drivers 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@295 -- # net_devs=() 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@295 -- # local -ga net_devs 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@296 -- # e810=() 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@296 -- # local -ga e810 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@297 -- # x722=() 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@297 -- # local -ga x722 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@298 -- # mlx=() 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@298 -- # local -ga mlx 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:22:44.534 Found 0000:86:00.0 (0x8086 - 0x159b) 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:22:44.534 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:22:44.535 Found 0000:86:00.1 (0x8086 - 0x159b) 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@390 -- # [[ up == up ]] 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:22:44.535 Found net devices under 0000:86:00.0: cvl_0_0 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@390 -- # [[ up == up ]] 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:22:44.535 Found net devices under 0000:86:00.1: cvl_0_1 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@414 -- # is_hw=yes 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:22:44.535 18:08:37 nvmf_tcp.nvmf_failover -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:22:44.535 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:22:44.535 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.184 ms 00:22:44.535 00:22:44.535 --- 10.0.0.2 ping statistics --- 00:22:44.535 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:44.535 rtt min/avg/max/mdev = 0.184/0.184/0.184/0.000 ms 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:22:44.535 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:22:44.535 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.102 ms 00:22:44.535 00:22:44.535 --- 10.0.0.1 ping statistics --- 00:22:44.535 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:22:44.535 rtt min/avg/max/mdev = 0.102/0.102/0.102/0.000 ms 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@422 -- # return 0 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- host/failover.sh@20 -- # nvmfappstart -m 0xE 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@722 -- # xtrace_disable 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@481 -- # nvmfpid=682471 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@482 -- # waitforlisten 682471 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@829 -- # '[' -z 682471 ']' 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@834 -- # local max_retries=100 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:22:44.535 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@838 -- # xtrace_disable 00:22:44.535 18:08:38 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:22:44.535 [2024-07-15 18:08:38.247845] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:22:44.535 [2024-07-15 18:08:38.247891] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:22:44.795 EAL: No free 2048 kB hugepages reported on node 1 00:22:44.795 [2024-07-15 18:08:38.304988] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:22:44.795 [2024-07-15 18:08:38.385732] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:22:44.795 [2024-07-15 18:08:38.385768] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:22:44.795 [2024-07-15 18:08:38.385775] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:22:44.795 [2024-07-15 18:08:38.385782] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:22:44.795 [2024-07-15 18:08:38.385787] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:22:44.795 [2024-07-15 18:08:38.386792] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:22:44.795 [2024-07-15 18:08:38.386877] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:22:44.795 [2024-07-15 18:08:38.386878] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:22:45.362 18:08:39 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:22:45.362 18:08:39 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@862 -- # return 0 00:22:45.362 18:08:39 nvmf_tcp.nvmf_failover -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:22:45.362 18:08:39 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@728 -- # xtrace_disable 00:22:45.362 18:08:39 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:22:45.620 18:08:39 nvmf_tcp.nvmf_failover -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:22:45.620 18:08:39 nvmf_tcp.nvmf_failover -- host/failover.sh@22 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:22:45.620 [2024-07-15 18:08:39.243909] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:22:45.620 18:08:39 nvmf_tcp.nvmf_failover -- host/failover.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc0 00:22:45.879 Malloc0 00:22:45.879 18:08:39 nvmf_tcp.nvmf_failover -- host/failover.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:22:46.138 18:08:39 nvmf_tcp.nvmf_failover -- host/failover.sh@25 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:22:46.138 18:08:39 nvmf_tcp.nvmf_failover -- host/failover.sh@26 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:22:46.397 [2024-07-15 18:08:40.001945] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:22:46.397 18:08:40 nvmf_tcp.nvmf_failover -- host/failover.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:22:46.655 [2024-07-15 18:08:40.186489] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:22:46.655 18:08:40 nvmf_tcp.nvmf_failover -- host/failover.sh@28 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4422 00:22:46.655 [2024-07-15 18:08:40.367059] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4422 *** 00:22:46.916 18:08:40 nvmf_tcp.nvmf_failover -- host/failover.sh@31 -- # bdevperf_pid=682903 00:22:46.916 18:08:40 nvmf_tcp.nvmf_failover -- host/failover.sh@30 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 15 -f 00:22:46.916 18:08:40 nvmf_tcp.nvmf_failover -- host/failover.sh@33 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; cat $testdir/try.txt; rm -f $testdir/try.txt; killprocess $bdevperf_pid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:22:46.916 18:08:40 nvmf_tcp.nvmf_failover -- host/failover.sh@34 -- # waitforlisten 682903 /var/tmp/bdevperf.sock 00:22:46.916 18:08:40 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@829 -- # '[' -z 682903 ']' 00:22:46.916 18:08:40 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:22:46.916 18:08:40 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@834 -- # local max_retries=100 00:22:46.917 18:08:40 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:22:46.917 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:22:46.917 18:08:40 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@838 -- # xtrace_disable 00:22:46.917 18:08:40 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:22:47.894 18:08:41 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:22:47.894 18:08:41 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@862 -- # return 0 00:22:47.895 18:08:41 nvmf_tcp.nvmf_failover -- host/failover.sh@35 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:22:47.895 NVMe0n1 00:22:48.152 18:08:41 nvmf_tcp.nvmf_failover -- host/failover.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:22:48.411 00:22:48.411 18:08:41 nvmf_tcp.nvmf_failover -- host/failover.sh@38 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:22:48.411 18:08:41 nvmf_tcp.nvmf_failover -- host/failover.sh@39 -- # run_test_pid=683147 00:22:48.411 18:08:41 nvmf_tcp.nvmf_failover -- host/failover.sh@41 -- # sleep 1 00:22:49.344 18:08:42 nvmf_tcp.nvmf_failover -- host/failover.sh@43 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:22:49.602 [2024-07-15 18:08:43.091836] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.602 [2024-07-15 18:08:43.091883] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.602 [2024-07-15 18:08:43.091890] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.602 [2024-07-15 18:08:43.091897] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.602 [2024-07-15 18:08:43.091903] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.602 [2024-07-15 18:08:43.091915] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.602 [2024-07-15 18:08:43.091921] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.602 [2024-07-15 18:08:43.091927] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.602 [2024-07-15 18:08:43.091933] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.602 [2024-07-15 18:08:43.091938] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.602 [2024-07-15 18:08:43.091944] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.602 [2024-07-15 18:08:43.091950] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.602 [2024-07-15 18:08:43.091956] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.091962] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.091967] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.091973] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.091978] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.091984] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.091990] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.091995] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092001] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092006] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092012] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092018] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092024] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092030] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092037] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092042] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092048] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092054] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092059] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092065] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092072] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092078] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092084] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092090] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092095] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092101] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092106] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092112] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092118] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092124] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092129] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092135] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092140] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092146] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092152] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092157] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092163] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092169] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092175] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092180] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092186] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092191] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092197] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092203] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092208] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092214] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092220] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092232] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092239] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092245] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092251] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092257] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092263] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092269] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092275] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092281] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092287] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092293] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092299] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092305] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092310] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092316] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092322] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092328] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092334] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092340] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092345] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092351] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092358] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092364] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092371] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092377] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092383] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092390] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 [2024-07-15 18:08:43.092398] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf2100 is same with the state(5) to be set 00:22:49.603 18:08:43 nvmf_tcp.nvmf_failover -- host/failover.sh@45 -- # sleep 3 00:22:52.888 18:08:46 nvmf_tcp.nvmf_failover -- host/failover.sh@47 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4422 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:22:52.888 00:22:52.888 18:08:46 nvmf_tcp.nvmf_failover -- host/failover.sh@48 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:22:52.888 [2024-07-15 18:08:46.605402] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605446] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605453] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605459] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605466] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605472] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605478] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605485] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605490] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605497] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605503] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605508] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605514] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605520] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605526] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605532] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605537] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605543] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605549] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605555] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605562] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605568] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605575] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605581] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605592] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605599] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605605] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605611] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605617] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605624] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:52.888 [2024-07-15 18:08:46.605630] tcp.c:1663:nvmf_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0xcf3660 is same with the state(5) to be set 00:22:53.146 18:08:46 nvmf_tcp.nvmf_failover -- host/failover.sh@50 -- # sleep 3 00:22:56.437 18:08:49 nvmf_tcp.nvmf_failover -- host/failover.sh@53 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:22:56.437 [2024-07-15 18:08:49.802803] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:22:56.437 18:08:49 nvmf_tcp.nvmf_failover -- host/failover.sh@55 -- # sleep 1 00:22:57.374 18:08:50 nvmf_tcp.nvmf_failover -- host/failover.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4422 00:22:57.374 18:08:51 nvmf_tcp.nvmf_failover -- host/failover.sh@59 -- # wait 683147 00:23:04.037 0 00:23:04.037 18:08:57 nvmf_tcp.nvmf_failover -- host/failover.sh@61 -- # killprocess 682903 00:23:04.037 18:08:57 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@948 -- # '[' -z 682903 ']' 00:23:04.037 18:08:57 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@952 -- # kill -0 682903 00:23:04.037 18:08:57 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@953 -- # uname 00:23:04.037 18:08:57 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:23:04.037 18:08:57 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 682903 00:23:04.037 18:08:57 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:23:04.037 18:08:57 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:23:04.037 18:08:57 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@966 -- # echo 'killing process with pid 682903' 00:23:04.038 killing process with pid 682903 00:23:04.038 18:08:57 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@967 -- # kill 682903 00:23:04.038 18:08:57 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@972 -- # wait 682903 00:23:04.038 18:08:57 nvmf_tcp.nvmf_failover -- host/failover.sh@63 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:23:04.038 [2024-07-15 18:08:40.441465] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:23:04.038 [2024-07-15 18:08:40.441515] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid682903 ] 00:23:04.038 EAL: No free 2048 kB hugepages reported on node 1 00:23:04.038 [2024-07-15 18:08:40.495216] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:04.038 [2024-07-15 18:08:40.570499] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:23:04.038 Running I/O for 15 seconds... 00:23:04.038 [2024-07-15 18:08:43.093796] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:95856 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.093832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.093848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:95864 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.093856] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.093865] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:95872 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.093872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.093881] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:95880 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.093888] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.093896] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:95888 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.093903] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.093911] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:95896 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.093917] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.093925] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:95904 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.093933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.093941] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:95912 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.093948] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.093956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:95920 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.093963] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.093971] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:95928 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.093981] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.093990] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:95936 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.093996] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:95944 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.094015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094024] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:95952 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.094030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:95960 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.094045] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094053] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:95968 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.094059] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:95976 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.094074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:95984 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.094089] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094097] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:95992 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.094103] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094112] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:96000 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.094118] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094126] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:96008 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.094133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094141] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:96016 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.094147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:96024 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.094161] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094169] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:96032 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.094176] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094184] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:96040 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.094192] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094200] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:96112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.038 [2024-07-15 18:08:43.094206] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094214] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:96120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.038 [2024-07-15 18:08:43.094222] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094236] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:96128 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.038 [2024-07-15 18:08:43.094243] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:96136 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.038 [2024-07-15 18:08:43.094258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094266] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:96144 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.038 [2024-07-15 18:08:43.094272] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094280] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:96152 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.038 [2024-07-15 18:08:43.094287] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094295] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:96160 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.038 [2024-07-15 18:08:43.094301] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094310] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:96168 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.038 [2024-07-15 18:08:43.094316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094326] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:96176 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.038 [2024-07-15 18:08:43.094334] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094344] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:96048 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.038 [2024-07-15 18:08:43.094352] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.038 [2024-07-15 18:08:43.094360] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:96184 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.038 [2024-07-15 18:08:43.094367] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094378] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:96192 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094384] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094394] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:105 nsid:1 lba:96200 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094400] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094408] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:96208 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094415] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094423] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:96216 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094429] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094437] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:96224 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094445] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:100 nsid:1 lba:96232 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094460] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094467] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:113 nsid:1 lba:96240 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094474] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094482] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:96248 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094489] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094497] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:96256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094503] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094511] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:96264 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094525] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:96272 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094531] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:96280 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094554] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:96288 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094568] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:96296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094586] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:96304 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094594] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094603] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:96312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094617] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:96320 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094624] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094631] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:96328 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094646] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:96336 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094652] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:96344 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094666] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094674] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:96352 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094681] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094689] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:96360 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094703] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:96056 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.039 [2024-07-15 18:08:43.094709] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:96064 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.039 [2024-07-15 18:08:43.094723] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094732] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:96072 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.039 [2024-07-15 18:08:43.094739] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094747] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:96080 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.039 [2024-07-15 18:08:43.094754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:96088 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.039 [2024-07-15 18:08:43.094770] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094778] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:96096 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.039 [2024-07-15 18:08:43.094785] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094793] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:96104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.039 [2024-07-15 18:08:43.094799] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094807] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:96368 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:118 nsid:1 lba:96376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094829] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094837] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:96384 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094851] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:96392 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094866] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:96400 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094881] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:96408 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094887] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094895] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:96416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094910] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:96424 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094916] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:96432 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094931] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094939] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:96440 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094945] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.039 [2024-07-15 18:08:43.094954] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:96448 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.039 [2024-07-15 18:08:43.094961] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.094969] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:96456 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.094976] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.094984] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:96464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.094990] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.094999] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:96472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095013] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:96480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:96488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:96496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095049] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095058] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:96504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095065] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095073] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:96512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095079] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:96520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095093] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095102] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:96528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095116] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:96536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095123] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:96544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095146] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:96552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095161] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:96560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095167] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095175] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:96568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095182] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095189] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:96576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095204] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:96584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095211] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095219] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:96592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095237] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:96600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095243] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095257] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:96608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095264] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095272] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:96616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095278] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095287] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:119 nsid:1 lba:96624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095293] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095301] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:96632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095308] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095316] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:96640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095322] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095332] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:96648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095338] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:96656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095353] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095360] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:96664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095367] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095375] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:96672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095381] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095389] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:96680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095395] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:96688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095417] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:96696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095424] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095432] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:96704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095438] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095446] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:96712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095453] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095460] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:96720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095467] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095475] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:96728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.040 [2024-07-15 18:08:43.095490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:96736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.040 [2024-07-15 18:08:43.095497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.095504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:96744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.041 [2024-07-15 18:08:43.095511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.095536] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.041 [2024-07-15 18:08:43.095542] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:96752 len:8 PRP1 0x0 PRP2 0x0 00:23:04.041 [2024-07-15 18:08:43.095549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.095558] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.041 [2024-07-15 18:08:43.095563] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.041 [2024-07-15 18:08:43.095568] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:96760 len:8 PRP1 0x0 PRP2 0x0 00:23:04.041 [2024-07-15 18:08:43.095575] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.095581] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.041 [2024-07-15 18:08:43.095586] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.041 [2024-07-15 18:08:43.095592] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:96768 len:8 PRP1 0x0 PRP2 0x0 00:23:04.041 [2024-07-15 18:08:43.095598] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.095604] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.041 [2024-07-15 18:08:43.095609] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.041 [2024-07-15 18:08:43.095615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:96776 len:8 PRP1 0x0 PRP2 0x0 00:23:04.041 [2024-07-15 18:08:43.095621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.095627] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.041 [2024-07-15 18:08:43.095632] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.041 [2024-07-15 18:08:43.095637] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:96784 len:8 PRP1 0x0 PRP2 0x0 00:23:04.041 [2024-07-15 18:08:43.095643] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.095650] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.041 [2024-07-15 18:08:43.095654] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.041 [2024-07-15 18:08:43.095660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:96792 len:8 PRP1 0x0 PRP2 0x0 00:23:04.041 [2024-07-15 18:08:43.095666] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.095672] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.041 [2024-07-15 18:08:43.095677] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.041 [2024-07-15 18:08:43.095682] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:96800 len:8 PRP1 0x0 PRP2 0x0 00:23:04.041 [2024-07-15 18:08:43.095688] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.095695] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.041 [2024-07-15 18:08:43.095702] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.041 [2024-07-15 18:08:43.095707] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:96808 len:8 PRP1 0x0 PRP2 0x0 00:23:04.041 [2024-07-15 18:08:43.095715] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.095722] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.041 [2024-07-15 18:08:43.095728] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.041 [2024-07-15 18:08:43.095734] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:96816 len:8 PRP1 0x0 PRP2 0x0 00:23:04.041 [2024-07-15 18:08:43.095740] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.095747] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.041 [2024-07-15 18:08:43.095752] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.041 [2024-07-15 18:08:43.095757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:96824 len:8 PRP1 0x0 PRP2 0x0 00:23:04.041 [2024-07-15 18:08:43.095764] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.095770] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.041 [2024-07-15 18:08:43.095775] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.041 [2024-07-15 18:08:43.095780] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:96832 len:8 PRP1 0x0 PRP2 0x0 00:23:04.041 [2024-07-15 18:08:43.095787] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.095794] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.041 [2024-07-15 18:08:43.095799] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.041 [2024-07-15 18:08:43.095804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:96840 len:8 PRP1 0x0 PRP2 0x0 00:23:04.041 [2024-07-15 18:08:43.095810] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.095817] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.041 [2024-07-15 18:08:43.095821] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.041 [2024-07-15 18:08:43.095827] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:96848 len:8 PRP1 0x0 PRP2 0x0 00:23:04.041 [2024-07-15 18:08:43.095833] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.095840] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.041 [2024-07-15 18:08:43.095845] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.041 [2024-07-15 18:08:43.095850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:96856 len:8 PRP1 0x0 PRP2 0x0 00:23:04.041 [2024-07-15 18:08:43.095856] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.107540] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.041 [2024-07-15 18:08:43.107550] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.041 [2024-07-15 18:08:43.107556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:96864 len:8 PRP1 0x0 PRP2 0x0 00:23:04.041 [2024-07-15 18:08:43.107564] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.107570] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.041 [2024-07-15 18:08:43.107576] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.041 [2024-07-15 18:08:43.107584] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:96872 len:8 PRP1 0x0 PRP2 0x0 00:23:04.041 [2024-07-15 18:08:43.107591] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.107631] bdev_nvme.c:1612:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x24e6300 was disconnected and freed. reset controller. 00:23:04.041 [2024-07-15 18:08:43.107640] bdev_nvme.c:1870:bdev_nvme_failover_trid: *NOTICE*: Start failover from 10.0.0.2:4420 to 10.0.0.2:4421 00:23:04.041 [2024-07-15 18:08:43.107661] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:23:04.041 [2024-07-15 18:08:43.107668] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.107675] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:23:04.041 [2024-07-15 18:08:43.107682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.107689] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:23:04.041 [2024-07-15 18:08:43.107696] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.107703] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:23:04.041 [2024-07-15 18:08:43.107709] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.041 [2024-07-15 18:08:43.107721] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:23:04.041 [2024-07-15 18:08:43.107750] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x24c8540 (9): Bad file descriptor 00:23:04.041 [2024-07-15 18:08:43.110898] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:23:04.041 [2024-07-15 18:08:43.145358] bdev_nvme.c:2067:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:23:04.041 [2024-07-15 18:08:46.605804] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:23:04.041 [2024-07-15 18:08:46.605839] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.605849] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:23:04.042 [2024-07-15 18:08:46.605857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.605865] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:23:04.042 [2024-07-15 18:08:46.605872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.605879] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:23:04.042 [2024-07-15 18:08:46.605885] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.605892] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x24c8540 is same with the state(5) to be set 00:23:04.042 [2024-07-15 18:08:46.605946] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:15496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.605954] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.605970] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:15504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.605978] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.605986] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:15512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.605993] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606001] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:15520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606009] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606017] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:15528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606024] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606032] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:15536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606041] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:15544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606057] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606068] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:15552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606083] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:15560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606090] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606098] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:15568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606104] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606112] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:118 nsid:1 lba:15576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606123] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606132] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:15584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606139] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:15592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606155] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606163] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:15600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606172] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606180] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:15608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606186] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:15616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606200] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606208] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:15624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606215] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606223] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:15632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606236] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606244] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:15640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606251] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606259] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:15648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606266] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606274] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:15656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:15664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606295] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606303] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:15672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606309] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.042 [2024-07-15 18:08:46.606319] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:15680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.042 [2024-07-15 18:08:46.606325] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606333] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:15688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606340] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606348] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:15696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606354] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606362] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:15704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606380] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:15712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606387] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:15720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606410] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:113 nsid:1 lba:15728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606417] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606425] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:15736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606432] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606440] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:15744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606447] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606454] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:15752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606461] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606469] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:15760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606484] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:15768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606491] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606500] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:15776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606506] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606514] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:15784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606520] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606528] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:15792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606535] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606543] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:15800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606558] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:15808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606565] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:15816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606587] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:15824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606594] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606602] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:15832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606616] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:15840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606623] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606631] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:15352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.043 [2024-07-15 18:08:46.606638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606653] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:15360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.043 [2024-07-15 18:08:46.606660] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606668] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:15848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606675] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606683] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:15856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606689] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606697] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:15864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606712] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:15872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606718] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606726] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:15880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606740] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:15888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:100 nsid:1 lba:15896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:15904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606785] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:15912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606792] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:15920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606806] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606814] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:90 nsid:1 lba:15928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606820] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606828] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:15936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606834] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606843] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:15944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.043 [2024-07-15 18:08:46.606849] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.043 [2024-07-15 18:08:46.606858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:15952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.606864] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.606872] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:15960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.606879] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.606889] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:15968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.606896] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.606904] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:15976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.606910] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.606918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:15984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.606925] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.606934] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:15992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.606941] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.606949] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:16000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.606955] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.606963] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:16008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.606969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.606978] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:16016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.606984] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.606992] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:16024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.606998] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607006] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:16032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607013] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607021] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:16040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607027] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607035] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:16048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:16056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607056] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607064] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:16064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607070] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:16072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607085] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607093] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:16080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607099] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607109] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:16088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607116] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607126] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:16096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607141] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:16104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:16112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607162] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607169] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:16120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607176] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607184] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:16128 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607198] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:16136 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607205] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607212] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:105 nsid:1 lba:16144 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607219] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607230] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:16152 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607237] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607245] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:16160 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607251] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607259] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16168 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607273] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:16176 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:16184 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607294] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607302] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:16192 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607310] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607318] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:16200 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607324] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607332] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:16208 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607338] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607348] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:16216 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607354] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607363] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:16224 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607369] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:16232 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607383] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607392] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:16240 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607398] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.044 [2024-07-15 18:08:46.607406] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:16248 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.044 [2024-07-15 18:08:46.607412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607420] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:16256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.045 [2024-07-15 18:08:46.607426] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:16264 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.045 [2024-07-15 18:08:46.607441] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607449] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:16272 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.045 [2024-07-15 18:08:46.607455] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607463] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:16280 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.045 [2024-07-15 18:08:46.607469] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607477] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:16288 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.045 [2024-07-15 18:08:46.607483] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607495] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:16296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.045 [2024-07-15 18:08:46.607501] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607509] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:16304 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.045 [2024-07-15 18:08:46.607515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:16312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.045 [2024-07-15 18:08:46.607530] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607538] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:16320 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.045 [2024-07-15 18:08:46.607544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607552] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:16328 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.045 [2024-07-15 18:08:46.607558] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607566] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:16336 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.045 [2024-07-15 18:08:46.607572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607582] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:16344 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.045 [2024-07-15 18:08:46.607588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607597] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:16352 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.045 [2024-07-15 18:08:46.607603] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607611] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:16360 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.045 [2024-07-15 18:08:46.607618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607627] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:15368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.045 [2024-07-15 18:08:46.607634] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:15376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.045 [2024-07-15 18:08:46.607649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607657] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:15384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.045 [2024-07-15 18:08:46.607663] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607671] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:15392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.045 [2024-07-15 18:08:46.607679] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607688] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:15400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.045 [2024-07-15 18:08:46.607694] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607702] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:15408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.045 [2024-07-15 18:08:46.607709] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:15416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.045 [2024-07-15 18:08:46.607723] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607732] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:16368 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.045 [2024-07-15 18:08:46.607739] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607748] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:15424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.045 [2024-07-15 18:08:46.607756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607765] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:15432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.045 [2024-07-15 18:08:46.607774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607783] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:66 nsid:1 lba:15440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.045 [2024-07-15 18:08:46.607790] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:15448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.045 [2024-07-15 18:08:46.607807] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:15456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.045 [2024-07-15 18:08:46.607821] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607830] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:15464 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.045 [2024-07-15 18:08:46.607836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607844] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:15472 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.045 [2024-07-15 18:08:46.607851] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607859] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:15480 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.045 [2024-07-15 18:08:46.607865] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607883] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.045 [2024-07-15 18:08:46.607890] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.045 [2024-07-15 18:08:46.607896] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:15488 len:8 PRP1 0x0 PRP2 0x0 00:23:04.045 [2024-07-15 18:08:46.607903] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:46.607944] bdev_nvme.c:1612:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x2693380 was disconnected and freed. reset controller. 00:23:04.045 [2024-07-15 18:08:46.607952] bdev_nvme.c:1870:bdev_nvme_failover_trid: *NOTICE*: Start failover from 10.0.0.2:4421 to 10.0.0.2:4422 00:23:04.045 [2024-07-15 18:08:46.607959] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:23:04.045 [2024-07-15 18:08:46.610776] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:23:04.045 [2024-07-15 18:08:46.610805] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x24c8540 (9): Bad file descriptor 00:23:04.045 [2024-07-15 18:08:46.728166] bdev_nvme.c:2067:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:23:04.045 [2024-07-15 18:08:50.986390] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:42872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.045 [2024-07-15 18:08:50.986439] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:50.986454] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:42880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.045 [2024-07-15 18:08:50.986462] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.045 [2024-07-15 18:08:50.986471] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:42888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.045 [2024-07-15 18:08:50.986478] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986486] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:42896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986493] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986501] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:42904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986516] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:42912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986522] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:90 nsid:1 lba:42920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986537] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986545] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:42928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986551] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986561] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:42936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986570] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986586] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:42944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986595] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986606] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:42952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986625] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:42960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986634] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986643] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:42968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986657] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:42976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986664] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986673] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:113 nsid:1 lba:42984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986679] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986687] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:42992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986701] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:43000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986708] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:43008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986724] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986731] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:43016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986738] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986746] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:43024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986753] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:43032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986775] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:43040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986783] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986792] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:43048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986799] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986807] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:43056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986823] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:43064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986829] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986840] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:43072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986848] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986860] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:43080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986867] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:43088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986891] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:43096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986897] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986905] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:43104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986912] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986920] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:118 nsid:1 lba:43112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986929] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986940] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:43120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986948] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:43128 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986963] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986971] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:43136 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986977] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.986987] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:43144 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.986994] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.987002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:43152 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.987008] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.987016] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:43160 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.987022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.987030] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:43168 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.987036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.987044] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:43176 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.987051] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.987059] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:43184 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.987065] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.987073] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:43192 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.046 [2024-07-15 18:08:50.987080] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.987088] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:42632 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.046 [2024-07-15 18:08:50.987095] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.987103] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:42640 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.046 [2024-07-15 18:08:50.987109] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.987117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:42648 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.046 [2024-07-15 18:08:50.987124] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.987132] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:42656 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.046 [2024-07-15 18:08:50.987138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.987146] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:42664 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.046 [2024-07-15 18:08:50.987152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.987160] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:42672 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.046 [2024-07-15 18:08:50.987168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.987176] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:42680 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.046 [2024-07-15 18:08:50.987183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.987191] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:42688 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.046 [2024-07-15 18:08:50.987198] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.987206] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:42696 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.046 [2024-07-15 18:08:50.987212] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.046 [2024-07-15 18:08:50.987220] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:42704 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.046 [2024-07-15 18:08:50.987233] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987243] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:42712 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.047 [2024-07-15 18:08:50.987251] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:42720 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.047 [2024-07-15 18:08:50.987270] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987280] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:42728 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.047 [2024-07-15 18:08:50.987289] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987297] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:42736 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.047 [2024-07-15 18:08:50.987304] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:42744 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.047 [2024-07-15 18:08:50.987318] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:42752 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.047 [2024-07-15 18:08:50.987333] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987342] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:42760 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.047 [2024-07-15 18:08:50.987348] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:42768 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.047 [2024-07-15 18:08:50.987363] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987370] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:42776 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.047 [2024-07-15 18:08:50.987378] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:42784 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.047 [2024-07-15 18:08:50.987393] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987401] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:42792 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.047 [2024-07-15 18:08:50.987408] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987415] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:42800 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:23:04.047 [2024-07-15 18:08:50.987422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987429] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:43200 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987436] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987444] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:43208 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987459] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:43216 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987465] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987473] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:43224 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:43232 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987501] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:43240 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987515] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:43248 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987522] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987529] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:43256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987536] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987544] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:43264 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987551] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987561] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:43272 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987567] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:43280 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:43288 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987596] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987604] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:43296 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987610] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987618] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:43304 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987624] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987632] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:43312 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987647] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:43320 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987653] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987661] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:43328 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987667] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987675] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:43336 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987690] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:43344 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987696] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987704] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:43352 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987711] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987718] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:43360 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987725] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987733] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:43368 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987740] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987749] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:105 nsid:1 lba:43376 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987755] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987763] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:119 nsid:1 lba:43384 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987770] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987778] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:43392 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987792] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:43400 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987798] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987806] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:43408 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987813] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987821] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:43416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987827] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987834] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:43424 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987849] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:43432 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987855] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987863] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:43440 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987869] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987877] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:43448 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987883] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987891] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:43456 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987898] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987905] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:43464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987912] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.047 [2024-07-15 18:08:50.987921] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:43472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.047 [2024-07-15 18:08:50.987928] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.987936] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:43480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.048 [2024-07-15 18:08:50.987942] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.987950] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:43488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.048 [2024-07-15 18:08:50.987957] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.987965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:43496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.048 [2024-07-15 18:08:50.987971] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.987979] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:43504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.048 [2024-07-15 18:08:50.987985] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.987993] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:43512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.048 [2024-07-15 18:08:50.988000] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988007] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:43520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.048 [2024-07-15 18:08:50.988014] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988022] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:43528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.048 [2024-07-15 18:08:50.988036] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988044] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:43536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.048 [2024-07-15 18:08:50.988051] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988059] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:43544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.048 [2024-07-15 18:08:50.988065] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988073] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:43552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.048 [2024-07-15 18:08:50.988079] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:43560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.048 [2024-07-15 18:08:50.988093] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988101] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:43568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.048 [2024-07-15 18:08:50.988108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:43576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:23:04.048 [2024-07-15 18:08:50.988123] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988143] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.048 [2024-07-15 18:08:50.988150] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:43584 len:8 PRP1 0x0 PRP2 0x0 00:23:04.048 [2024-07-15 18:08:50.988156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988165] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.048 [2024-07-15 18:08:50.988170] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.048 [2024-07-15 18:08:50.988175] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:43592 len:8 PRP1 0x0 PRP2 0x0 00:23:04.048 [2024-07-15 18:08:50.988181] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988188] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.048 [2024-07-15 18:08:50.988193] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.048 [2024-07-15 18:08:50.988198] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:43600 len:8 PRP1 0x0 PRP2 0x0 00:23:04.048 [2024-07-15 18:08:50.988205] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988211] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.048 [2024-07-15 18:08:50.988216] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.048 [2024-07-15 18:08:50.988221] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:43608 len:8 PRP1 0x0 PRP2 0x0 00:23:04.048 [2024-07-15 18:08:50.988233] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988240] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.048 [2024-07-15 18:08:50.988245] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.048 [2024-07-15 18:08:50.988250] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:43616 len:8 PRP1 0x0 PRP2 0x0 00:23:04.048 [2024-07-15 18:08:50.988256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988265] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.048 [2024-07-15 18:08:50.988270] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.048 [2024-07-15 18:08:50.988275] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:43624 len:8 PRP1 0x0 PRP2 0x0 00:23:04.048 [2024-07-15 18:08:50.988281] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988288] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.048 [2024-07-15 18:08:50.988293] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.048 [2024-07-15 18:08:50.988298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:43632 len:8 PRP1 0x0 PRP2 0x0 00:23:04.048 [2024-07-15 18:08:50.988304] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988311] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.048 [2024-07-15 18:08:50.988317] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.048 [2024-07-15 18:08:50.988323] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:43640 len:8 PRP1 0x0 PRP2 0x0 00:23:04.048 [2024-07-15 18:08:50.988329] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988335] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.048 [2024-07-15 18:08:50.988340] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.048 [2024-07-15 18:08:50.988345] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:42808 len:8 PRP1 0x0 PRP2 0x0 00:23:04.048 [2024-07-15 18:08:50.988351] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988358] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.048 [2024-07-15 18:08:50.988363] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.048 [2024-07-15 18:08:50.988369] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:42816 len:8 PRP1 0x0 PRP2 0x0 00:23:04.048 [2024-07-15 18:08:50.988375] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988381] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.048 [2024-07-15 18:08:50.988386] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.048 [2024-07-15 18:08:50.988391] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:42824 len:8 PRP1 0x0 PRP2 0x0 00:23:04.048 [2024-07-15 18:08:50.988397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988403] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.048 [2024-07-15 18:08:50.988409] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.048 [2024-07-15 18:08:50.988414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:42832 len:8 PRP1 0x0 PRP2 0x0 00:23:04.048 [2024-07-15 18:08:50.988420] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988426] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.048 [2024-07-15 18:08:50.988431] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.048 [2024-07-15 18:08:50.988436] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:42840 len:8 PRP1 0x0 PRP2 0x0 00:23:04.048 [2024-07-15 18:08:50.988442] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988450] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.048 [2024-07-15 18:08:50.988455] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.048 [2024-07-15 18:08:50.988460] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:42848 len:8 PRP1 0x0 PRP2 0x0 00:23:04.048 [2024-07-15 18:08:50.988466] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988473] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.048 [2024-07-15 18:08:50.988478] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.048 [2024-07-15 18:08:50.988483] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:42856 len:8 PRP1 0x0 PRP2 0x0 00:23:04.048 [2024-07-15 18:08:50.988489] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988497] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.048 [2024-07-15 18:08:50.988502] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.048 [2024-07-15 18:08:50.988507] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:42864 len:8 PRP1 0x0 PRP2 0x0 00:23:04.048 [2024-07-15 18:08:50.988513] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988520] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:23:04.048 [2024-07-15 18:08:50.988525] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:23:04.048 [2024-07-15 18:08:50.988530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:43648 len:8 PRP1 0x0 PRP2 0x0 00:23:04.048 [2024-07-15 18:08:50.988536] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988579] bdev_nvme.c:1612:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x2693170 was disconnected and freed. reset controller. 00:23:04.048 [2024-07-15 18:08:50.988588] bdev_nvme.c:1870:bdev_nvme_failover_trid: *NOTICE*: Start failover from 10.0.0.2:4422 to 10.0.0.2:4420 00:23:04.048 [2024-07-15 18:08:50.988609] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:23:04.048 [2024-07-15 18:08:50.988616] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988623] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:23:04.048 [2024-07-15 18:08:50.988630] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988637] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:23:04.048 [2024-07-15 18:08:50.988643] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988650] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:23:04.048 [2024-07-15 18:08:50.988656] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:04.048 [2024-07-15 18:08:50.988663] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:23:04.048 [2024-07-15 18:08:50.988684] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x24c8540 (9): Bad file descriptor 00:23:04.048 [2024-07-15 18:08:50.991505] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:23:04.048 [2024-07-15 18:08:51.067411] bdev_nvme.c:2067:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:23:04.048 00:23:04.048 Latency(us) 00:23:04.048 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:23:04.048 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:23:04.048 Verification LBA range: start 0x0 length 0x4000 00:23:04.048 NVMe0n1 : 15.01 10840.53 42.35 686.22 0.00 11082.38 594.81 21313.45 00:23:04.048 =================================================================================================================== 00:23:04.048 Total : 10840.53 42.35 686.22 0.00 11082.38 594.81 21313.45 00:23:04.048 Received shutdown signal, test time was about 15.000000 seconds 00:23:04.048 00:23:04.048 Latency(us) 00:23:04.048 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:23:04.048 =================================================================================================================== 00:23:04.048 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:23:04.049 18:08:57 nvmf_tcp.nvmf_failover -- host/failover.sh@65 -- # grep -c 'Resetting controller successful' 00:23:04.049 18:08:57 nvmf_tcp.nvmf_failover -- host/failover.sh@65 -- # count=3 00:23:04.049 18:08:57 nvmf_tcp.nvmf_failover -- host/failover.sh@67 -- # (( count != 3 )) 00:23:04.049 18:08:57 nvmf_tcp.nvmf_failover -- host/failover.sh@73 -- # bdevperf_pid=685664 00:23:04.049 18:08:57 nvmf_tcp.nvmf_failover -- host/failover.sh@72 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 1 -f 00:23:04.049 18:08:57 nvmf_tcp.nvmf_failover -- host/failover.sh@75 -- # waitforlisten 685664 /var/tmp/bdevperf.sock 00:23:04.049 18:08:57 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@829 -- # '[' -z 685664 ']' 00:23:04.049 18:08:57 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:23:04.049 18:08:57 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@834 -- # local max_retries=100 00:23:04.049 18:08:57 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:23:04.049 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:23:04.049 18:08:57 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@838 -- # xtrace_disable 00:23:04.049 18:08:57 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:23:04.615 18:08:58 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:23:04.615 18:08:58 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@862 -- # return 0 00:23:04.615 18:08:58 nvmf_tcp.nvmf_failover -- host/failover.sh@76 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:23:04.615 [2024-07-15 18:08:58.326681] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:23:04.873 18:08:58 nvmf_tcp.nvmf_failover -- host/failover.sh@77 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4422 00:23:04.873 [2024-07-15 18:08:58.499138] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4422 *** 00:23:04.873 18:08:58 nvmf_tcp.nvmf_failover -- host/failover.sh@78 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:23:05.440 NVMe0n1 00:23:05.440 18:08:58 nvmf_tcp.nvmf_failover -- host/failover.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:23:05.440 00:23:05.698 18:08:59 nvmf_tcp.nvmf_failover -- host/failover.sh@80 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b NVMe0 -t tcp -a 10.0.0.2 -s 4422 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:23:05.957 00:23:05.957 18:08:59 nvmf_tcp.nvmf_failover -- host/failover.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:23:05.957 18:08:59 nvmf_tcp.nvmf_failover -- host/failover.sh@82 -- # grep -q NVMe0 00:23:06.215 18:08:59 nvmf_tcp.nvmf_failover -- host/failover.sh@84 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:23:06.215 18:08:59 nvmf_tcp.nvmf_failover -- host/failover.sh@87 -- # sleep 3 00:23:09.494 18:09:02 nvmf_tcp.nvmf_failover -- host/failover.sh@88 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:23:09.494 18:09:02 nvmf_tcp.nvmf_failover -- host/failover.sh@88 -- # grep -q NVMe0 00:23:09.494 18:09:03 nvmf_tcp.nvmf_failover -- host/failover.sh@90 -- # run_test_pid=686720 00:23:09.494 18:09:03 nvmf_tcp.nvmf_failover -- host/failover.sh@89 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bdevperf.sock perform_tests 00:23:09.494 18:09:03 nvmf_tcp.nvmf_failover -- host/failover.sh@92 -- # wait 686720 00:23:10.903 0 00:23:10.903 18:09:04 nvmf_tcp.nvmf_failover -- host/failover.sh@94 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:23:10.903 [2024-07-15 18:08:57.346706] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:23:10.903 [2024-07-15 18:08:57.346755] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid685664 ] 00:23:10.903 EAL: No free 2048 kB hugepages reported on node 1 00:23:10.903 [2024-07-15 18:08:57.400245] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:10.903 [2024-07-15 18:08:57.469174] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:23:10.903 [2024-07-15 18:08:59.885842] bdev_nvme.c:1870:bdev_nvme_failover_trid: *NOTICE*: Start failover from 10.0.0.2:4420 to 10.0.0.2:4421 00:23:10.903 [2024-07-15 18:08:59.885888] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:23:10.903 [2024-07-15 18:08:59.885899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:10.903 [2024-07-15 18:08:59.885908] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:23:10.903 [2024-07-15 18:08:59.885914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:10.903 [2024-07-15 18:08:59.885921] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:23:10.903 [2024-07-15 18:08:59.885928] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:10.903 [2024-07-15 18:08:59.885935] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:23:10.903 [2024-07-15 18:08:59.885941] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:10.903 [2024-07-15 18:08:59.885952] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:23:10.903 [2024-07-15 18:08:59.885975] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:23:10.903 [2024-07-15 18:08:59.885987] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x8e1540 (9): Bad file descriptor 00:23:10.903 [2024-07-15 18:08:59.897912] bdev_nvme.c:2067:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:23:10.903 Running I/O for 1 seconds... 00:23:10.903 00:23:10.903 Latency(us) 00:23:10.903 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:23:10.903 Job: NVMe0n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:23:10.903 Verification LBA range: start 0x0 length 0x4000 00:23:10.903 NVMe0n1 : 1.01 11120.11 43.44 0.00 0.00 11464.38 2251.02 11625.52 00:23:10.903 =================================================================================================================== 00:23:10.903 Total : 11120.11 43.44 0.00 0.00 11464.38 2251.02 11625.52 00:23:10.903 18:09:04 nvmf_tcp.nvmf_failover -- host/failover.sh@95 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:23:10.903 18:09:04 nvmf_tcp.nvmf_failover -- host/failover.sh@95 -- # grep -q NVMe0 00:23:10.903 18:09:04 nvmf_tcp.nvmf_failover -- host/failover.sh@98 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4422 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:23:10.903 18:09:04 nvmf_tcp.nvmf_failover -- host/failover.sh@99 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:23:10.903 18:09:04 nvmf_tcp.nvmf_failover -- host/failover.sh@99 -- # grep -q NVMe0 00:23:11.162 18:09:04 nvmf_tcp.nvmf_failover -- host/failover.sh@100 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_detach_controller NVMe0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 00:23:11.419 18:09:04 nvmf_tcp.nvmf_failover -- host/failover.sh@101 -- # sleep 3 00:23:14.701 18:09:07 nvmf_tcp.nvmf_failover -- host/failover.sh@103 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_controllers 00:23:14.701 18:09:07 nvmf_tcp.nvmf_failover -- host/failover.sh@103 -- # grep -q NVMe0 00:23:14.701 18:09:08 nvmf_tcp.nvmf_failover -- host/failover.sh@108 -- # killprocess 685664 00:23:14.701 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@948 -- # '[' -z 685664 ']' 00:23:14.701 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@952 -- # kill -0 685664 00:23:14.701 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@953 -- # uname 00:23:14.701 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:23:14.701 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 685664 00:23:14.701 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:23:14.701 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:23:14.701 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@966 -- # echo 'killing process with pid 685664' 00:23:14.701 killing process with pid 685664 00:23:14.701 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@967 -- # kill 685664 00:23:14.701 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@972 -- # wait 685664 00:23:14.701 18:09:08 nvmf_tcp.nvmf_failover -- host/failover.sh@110 -- # sync 00:23:14.701 18:09:08 nvmf_tcp.nvmf_failover -- host/failover.sh@111 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:23:14.960 18:09:08 nvmf_tcp.nvmf_failover -- host/failover.sh@113 -- # trap - SIGINT SIGTERM EXIT 00:23:14.960 18:09:08 nvmf_tcp.nvmf_failover -- host/failover.sh@115 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:23:14.960 18:09:08 nvmf_tcp.nvmf_failover -- host/failover.sh@116 -- # nvmftestfini 00:23:14.960 18:09:08 nvmf_tcp.nvmf_failover -- nvmf/common.sh@488 -- # nvmfcleanup 00:23:14.960 18:09:08 nvmf_tcp.nvmf_failover -- nvmf/common.sh@117 -- # sync 00:23:14.960 18:09:08 nvmf_tcp.nvmf_failover -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:23:14.960 18:09:08 nvmf_tcp.nvmf_failover -- nvmf/common.sh@120 -- # set +e 00:23:14.960 18:09:08 nvmf_tcp.nvmf_failover -- nvmf/common.sh@121 -- # for i in {1..20} 00:23:14.960 18:09:08 nvmf_tcp.nvmf_failover -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:23:14.960 rmmod nvme_tcp 00:23:14.960 rmmod nvme_fabrics 00:23:14.960 rmmod nvme_keyring 00:23:14.960 18:09:08 nvmf_tcp.nvmf_failover -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:23:14.960 18:09:08 nvmf_tcp.nvmf_failover -- nvmf/common.sh@124 -- # set -e 00:23:14.960 18:09:08 nvmf_tcp.nvmf_failover -- nvmf/common.sh@125 -- # return 0 00:23:14.960 18:09:08 nvmf_tcp.nvmf_failover -- nvmf/common.sh@489 -- # '[' -n 682471 ']' 00:23:14.960 18:09:08 nvmf_tcp.nvmf_failover -- nvmf/common.sh@490 -- # killprocess 682471 00:23:14.960 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@948 -- # '[' -z 682471 ']' 00:23:14.960 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@952 -- # kill -0 682471 00:23:14.960 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@953 -- # uname 00:23:14.960 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:23:14.960 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 682471 00:23:15.219 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:23:15.219 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:23:15.219 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@966 -- # echo 'killing process with pid 682471' 00:23:15.219 killing process with pid 682471 00:23:15.219 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@967 -- # kill 682471 00:23:15.219 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@972 -- # wait 682471 00:23:15.219 18:09:08 nvmf_tcp.nvmf_failover -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:23:15.219 18:09:08 nvmf_tcp.nvmf_failover -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:23:15.219 18:09:08 nvmf_tcp.nvmf_failover -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:23:15.219 18:09:08 nvmf_tcp.nvmf_failover -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:23:15.219 18:09:08 nvmf_tcp.nvmf_failover -- nvmf/common.sh@278 -- # remove_spdk_ns 00:23:15.219 18:09:08 nvmf_tcp.nvmf_failover -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:15.219 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:23:15.219 18:09:08 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:17.753 18:09:10 nvmf_tcp.nvmf_failover -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:23:17.753 00:23:17.753 real 0m38.186s 00:23:17.753 user 2m3.744s 00:23:17.753 sys 0m7.247s 00:23:17.753 18:09:10 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@1124 -- # xtrace_disable 00:23:17.753 18:09:10 nvmf_tcp.nvmf_failover -- common/autotest_common.sh@10 -- # set +x 00:23:17.753 ************************************ 00:23:17.753 END TEST nvmf_failover 00:23:17.753 ************************************ 00:23:17.753 18:09:11 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:23:17.753 18:09:11 nvmf_tcp -- nvmf/nvmf.sh@101 -- # run_test nvmf_host_discovery /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery.sh --transport=tcp 00:23:17.753 18:09:11 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:23:17.753 18:09:11 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:23:17.753 18:09:11 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:23:17.753 ************************************ 00:23:17.753 START TEST nvmf_host_discovery 00:23:17.753 ************************************ 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery.sh --transport=tcp 00:23:17.753 * Looking for test storage... 00:23:17.753 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@7 -- # uname -s 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- paths/export.sh@5 -- # export PATH 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@47 -- # : 0 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:23:17.753 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@51 -- # have_pci_nics=0 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@11 -- # '[' tcp == rdma ']' 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@16 -- # DISCOVERY_PORT=8009 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@17 -- # DISCOVERY_NQN=nqn.2014-08.org.nvmexpress.discovery 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@20 -- # NQN=nqn.2016-06.io.spdk:cnode 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@22 -- # HOST_NQN=nqn.2021-12.io.spdk:test 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@23 -- # HOST_SOCK=/tmp/host.sock 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@25 -- # nvmftestinit 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@448 -- # prepare_net_devs 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@410 -- # local -g is_hw=no 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@412 -- # remove_spdk_ns 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@285 -- # xtrace_disable 00:23:17.754 18:09:11 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@291 -- # pci_devs=() 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@291 -- # local -a pci_devs 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@292 -- # pci_net_devs=() 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@293 -- # pci_drivers=() 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@293 -- # local -A pci_drivers 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@295 -- # net_devs=() 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@295 -- # local -ga net_devs 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@296 -- # e810=() 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@296 -- # local -ga e810 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@297 -- # x722=() 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@297 -- # local -ga x722 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@298 -- # mlx=() 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@298 -- # local -ga mlx 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:23:23.029 Found 0000:86:00.0 (0x8086 - 0x159b) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:23:23.029 Found 0000:86:00.1 (0x8086 - 0x159b) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@390 -- # [[ up == up ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:23:23.029 Found net devices under 0000:86:00.0: cvl_0_0 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@390 -- # [[ up == up ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:23:23.029 Found net devices under 0000:86:00.1: cvl_0_1 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@414 -- # is_hw=yes 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:23:23.029 18:09:15 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:23:23.029 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:23:23.029 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:23:23.029 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:23:23.029 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:23:23.029 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:23:23.029 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:23:23.029 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:23:23.029 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:23:23.029 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.175 ms 00:23:23.029 00:23:23.029 --- 10.0.0.2 ping statistics --- 00:23:23.029 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:23.029 rtt min/avg/max/mdev = 0.175/0.175/0.175/0.000 ms 00:23:23.029 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:23:23.029 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:23:23.029 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.247 ms 00:23:23.029 00:23:23.029 --- 10.0.0.1 ping statistics --- 00:23:23.029 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:23.029 rtt min/avg/max/mdev = 0.247/0.247/0.247/0.000 ms 00:23:23.029 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:23:23.029 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@422 -- # return 0 00:23:23.029 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:23:23.029 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:23:23.029 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:23:23.029 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:23:23.029 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:23:23.029 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:23:23.029 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:23:23.029 18:09:16 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@30 -- # nvmfappstart -m 0x2 00:23:23.030 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:23:23.030 18:09:16 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@722 -- # xtrace_disable 00:23:23.030 18:09:16 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:23.030 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@481 -- # nvmfpid=691324 00:23:23.030 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@482 -- # waitforlisten 691324 00:23:23.030 18:09:16 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:23:23.030 18:09:16 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@829 -- # '[' -z 691324 ']' 00:23:23.030 18:09:16 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:23.030 18:09:16 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@834 -- # local max_retries=100 00:23:23.030 18:09:16 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:23.030 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:23.030 18:09:16 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@838 -- # xtrace_disable 00:23:23.030 18:09:16 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:23.030 [2024-07-15 18:09:16.252524] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:23:23.030 [2024-07-15 18:09:16.252569] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:23:23.030 EAL: No free 2048 kB hugepages reported on node 1 00:23:23.030 [2024-07-15 18:09:16.308567] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:23.030 [2024-07-15 18:09:16.387079] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:23:23.030 [2024-07-15 18:09:16.387114] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:23:23.030 [2024-07-15 18:09:16.387121] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:23:23.030 [2024-07-15 18:09:16.387127] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:23:23.030 [2024-07-15 18:09:16.387132] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:23:23.030 [2024-07-15 18:09:16.387147] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@862 -- # return 0 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@728 -- # xtrace_disable 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@32 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:23.596 [2024-07-15 18:09:17.085425] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@33 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2014-08.org.nvmexpress.discovery -t tcp -a 10.0.0.2 -s 8009 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:23.596 [2024-07-15 18:09:17.093540] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 8009 *** 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@35 -- # rpc_cmd bdev_null_create null0 1000 512 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:23.596 null0 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@36 -- # rpc_cmd bdev_null_create null1 1000 512 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:23.596 null1 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@37 -- # rpc_cmd bdev_wait_for_examine 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@45 -- # hostpid=691568 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -m 0x1 -r /tmp/host.sock 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@46 -- # waitforlisten 691568 /tmp/host.sock 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@829 -- # '[' -z 691568 ']' 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@833 -- # local rpc_addr=/tmp/host.sock 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@834 -- # local max_retries=100 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock...' 00:23:23.596 Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock... 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@838 -- # xtrace_disable 00:23:23.596 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:23.596 [2024-07-15 18:09:17.168993] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:23:23.596 [2024-07-15 18:09:17.169032] [ DPDK EAL parameters: nvmf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid691568 ] 00:23:23.596 EAL: No free 2048 kB hugepages reported on node 1 00:23:23.596 [2024-07-15 18:09:17.223378] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:23:23.596 [2024-07-15 18:09:17.303241] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:23:24.529 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:23:24.529 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@862 -- # return 0 00:23:24.529 18:09:17 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@48 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; kill $hostpid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:23:24.529 18:09:17 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@50 -- # rpc_cmd -s /tmp/host.sock log_set_flag bdev_nvme 00:23:24.529 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:24.529 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:24.529 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:24.529 18:09:17 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@51 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test 00:23:24.529 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:24.529 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:24.529 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:24.529 18:09:17 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@72 -- # notify_id=0 00:23:24.529 18:09:17 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@83 -- # get_subsystem_names 00:23:24.529 18:09:17 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:23:24.529 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:24.529 18:09:17 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:24.529 18:09:17 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:23:24.529 18:09:17 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:23:24.529 18:09:17 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@83 -- # [[ '' == '' ]] 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@84 -- # get_bdev_list 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@84 -- # [[ '' == '' ]] 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@86 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@87 -- # get_subsystem_names 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@87 -- # [[ '' == '' ]] 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@88 -- # get_bdev_list 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@88 -- # [[ '' == '' ]] 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@90 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 null0 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@91 -- # get_subsystem_names 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:23:24.529 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@91 -- # [[ '' == '' ]] 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@92 -- # get_bdev_list 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@92 -- # [[ '' == '' ]] 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@96 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:24.787 [2024-07-15 18:09:18.312771] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@97 -- # get_subsystem_names 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@97 -- # [[ '' == '' ]] 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@98 -- # get_bdev_list 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@98 -- # [[ '' == '' ]] 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@99 -- # is_notification_count_eq 0 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=0 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@912 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@913 -- # local max=10 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@914 -- # (( max-- )) 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # get_notification_count 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 0 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=0 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=0 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # (( notification_count == expected_count )) 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@916 -- # return 0 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@103 -- # rpc_cmd nvmf_subsystem_add_host nqn.2016-06.io.spdk:cnode0 nqn.2021-12.io.spdk:test 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@105 -- # waitforcondition '[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@912 -- # local 'cond=[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@913 -- # local max=10 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@914 -- # (( max-- )) 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # get_subsystem_names 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:23:24.787 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:24.788 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # [[ '' == \n\v\m\e\0 ]] 00:23:24.788 18:09:18 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@918 -- # sleep 1 00:23:25.353 [2024-07-15 18:09:18.993108] bdev_nvme.c:6983:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:23:25.353 [2024-07-15 18:09:18.993127] bdev_nvme.c:7063:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:23:25.353 [2024-07-15 18:09:18.993138] bdev_nvme.c:6946:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:23:25.609 [2024-07-15 18:09:19.081423] bdev_nvme.c:6912:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 new subsystem nvme0 00:23:25.609 [2024-07-15 18:09:19.266753] bdev_nvme.c:6802:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:23:25.609 [2024-07-15 18:09:19.266772] bdev_nvme.c:6761:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@914 -- # (( max-- )) 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # get_subsystem_names 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@916 -- # return 0 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@106 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1" ]]' 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@912 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1" ]]' 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@913 -- # local max=10 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@914 -- # (( max-- )) 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1"' ']]' 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # get_bdev_list 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:25.866 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # [[ nvme0n1 == \n\v\m\e\0\n\1 ]] 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@916 -- # return 0 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@107 -- # waitforcondition '[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT" ]]' 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@912 -- # local 'cond=[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT" ]]' 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@913 -- # local max=10 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@914 -- # (( max-- )) 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_PORT"' ']]' 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # get_subsystem_paths nvme0 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@63 -- # xargs 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@63 -- # sort -n 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # [[ 4420 == \4\4\2\0 ]] 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@916 -- # return 0 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@108 -- # is_notification_count_eq 1 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=1 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@912 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@913 -- # local max=10 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@914 -- # (( max-- )) 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # get_notification_count 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 0 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=1 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=1 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # (( notification_count == expected_count )) 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@916 -- # return 0 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@111 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 null1 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@113 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@912 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@913 -- # local max=10 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@914 -- # (( max-- )) 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1' 'nvme0n2"' ']]' 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # get_bdev_list 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@916 -- # return 0 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@114 -- # is_notification_count_eq 1 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=1 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@912 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@913 -- # local max=10 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@914 -- # (( max-- )) 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # get_notification_count 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 1 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=1 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=2 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # (( notification_count == expected_count )) 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@916 -- # return 0 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@118 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4421 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:26.125 [2024-07-15 18:09:19.812938] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:23:26.125 [2024-07-15 18:09:19.813135] bdev_nvme.c:6965:discovery_aer_cb: *INFO*: Discovery[10.0.0.2:8009] got aer 00:23:26.125 [2024-07-15 18:09:19.813156] bdev_nvme.c:6946:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@120 -- # waitforcondition '[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@912 -- # local 'cond=[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@913 -- # local max=10 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@914 -- # (( max-- )) 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # get_subsystem_names 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:26.125 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@916 -- # return 0 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@121 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@912 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@913 -- # local max=10 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@914 -- # (( max-- )) 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1' 'nvme0n2"' ']]' 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # get_bdev_list 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:26.382 [2024-07-15 18:09:19.900419] bdev_nvme.c:6907:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 new path for nvme0 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@916 -- # return 0 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@122 -- # waitforcondition '[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT $NVMF_SECOND_PORT" ]]' 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@912 -- # local 'cond=[[ "$(get_subsystem_paths nvme0)" == "$NVMF_PORT $NVMF_SECOND_PORT" ]]' 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@913 -- # local max=10 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@914 -- # (( max-- )) 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_PORT' '$NVMF_SECOND_PORT"' ']]' 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # get_subsystem_paths nvme0 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@63 -- # xargs 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@63 -- # sort -n 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:26.382 [2024-07-15 18:09:19.958789] bdev_nvme.c:6802:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:23:26.382 [2024-07-15 18:09:19.958804] bdev_nvme.c:6761:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:23:26.382 [2024-07-15 18:09:19.958809] bdev_nvme.c:6761:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 found again 00:23:26.382 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # [[ 4420 == \4\4\2\0\ \4\4\2\1 ]] 00:23:26.383 18:09:19 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@918 -- # sleep 1 00:23:27.316 18:09:20 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@914 -- # (( max-- )) 00:23:27.316 18:09:20 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_PORT' '$NVMF_SECOND_PORT"' ']]' 00:23:27.316 18:09:20 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # get_subsystem_paths nvme0 00:23:27.316 18:09:20 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:23:27.316 18:09:20 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:23:27.316 18:09:20 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:27.316 18:09:20 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@63 -- # sort -n 00:23:27.316 18:09:20 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:27.316 18:09:20 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@63 -- # xargs 00:23:27.316 18:09:20 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:27.316 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # [[ 4420 4421 == \4\4\2\0\ \4\4\2\1 ]] 00:23:27.316 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@916 -- # return 0 00:23:27.316 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@123 -- # is_notification_count_eq 0 00:23:27.316 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=0 00:23:27.316 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:23:27.316 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@912 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:23:27.316 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@913 -- # local max=10 00:23:27.316 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@914 -- # (( max-- )) 00:23:27.316 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:23:27.316 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # get_notification_count 00:23:27.316 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 2 00:23:27.316 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:23:27.316 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:27.316 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:27.316 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=0 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=2 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # (( notification_count == expected_count )) 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@916 -- # return 0 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@127 -- # rpc_cmd nvmf_subsystem_remove_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:27.575 [2024-07-15 18:09:21.068919] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:23:27.575 [2024-07-15 18:09:21.068945] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:27.575 [2024-07-15 18:09:21.068955] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:23:27.575 [2024-07-15 18:09:21.068962] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:27.575 [2024-07-15 18:09:21.068969] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:23:27.575 [2024-07-15 18:09:21.068976] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:27.575 [2024-07-15 18:09:21.068983] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:23:27.575 [2024-07-15 18:09:21.068989] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:23:27.575 [2024-07-15 18:09:21.068995] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x668f10 is same with the state(5) to be set 00:23:27.575 [2024-07-15 18:09:21.069292] bdev_nvme.c:6965:discovery_aer_cb: *INFO*: Discovery[10.0.0.2:8009] got aer 00:23:27.575 [2024-07-15 18:09:21.069305] bdev_nvme.c:6946:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@129 -- # waitforcondition '[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@912 -- # local 'cond=[[ "$(get_subsystem_names)" == "nvme0" ]]' 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@913 -- # local max=10 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@914 -- # (( max-- )) 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # eval '[[' '"$(get_subsystem_names)"' == '"nvme0"' ']]' 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # get_subsystem_names 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:23:27.575 [2024-07-15 18:09:21.078928] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x668f10 (9): Bad file descriptor 00:23:27.575 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:27.575 [2024-07-15 18:09:21.088965] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:23:27.575 [2024-07-15 18:09:21.089282] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:27.575 [2024-07-15 18:09:21.089297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x668f10 with addr=10.0.0.2, port=4420 00:23:27.575 [2024-07-15 18:09:21.089305] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x668f10 is same with the state(5) to be set 00:23:27.575 [2024-07-15 18:09:21.089316] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x668f10 (9): Bad file descriptor 00:23:27.575 [2024-07-15 18:09:21.089327] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:23:27.575 [2024-07-15 18:09:21.089334] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:23:27.575 [2024-07-15 18:09:21.089341] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:23:27.575 [2024-07-15 18:09:21.089358] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:23:27.575 [2024-07-15 18:09:21.099020] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:23:27.575 [2024-07-15 18:09:21.099230] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:27.575 [2024-07-15 18:09:21.099242] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x668f10 with addr=10.0.0.2, port=4420 00:23:27.575 [2024-07-15 18:09:21.099249] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x668f10 is same with the state(5) to be set 00:23:27.575 [2024-07-15 18:09:21.099259] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x668f10 (9): Bad file descriptor 00:23:27.576 [2024-07-15 18:09:21.099268] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:23:27.576 [2024-07-15 18:09:21.099274] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:23:27.576 [2024-07-15 18:09:21.099281] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:23:27.576 [2024-07-15 18:09:21.099290] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:23:27.576 [2024-07-15 18:09:21.109066] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:23:27.576 [2024-07-15 18:09:21.109353] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:27.576 [2024-07-15 18:09:21.109365] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x668f10 with addr=10.0.0.2, port=4420 00:23:27.576 [2024-07-15 18:09:21.109372] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x668f10 is same with the state(5) to be set 00:23:27.576 [2024-07-15 18:09:21.109382] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x668f10 (9): Bad file descriptor 00:23:27.576 [2024-07-15 18:09:21.109391] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:23:27.576 [2024-07-15 18:09:21.109396] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:23:27.576 [2024-07-15 18:09:21.109403] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:23:27.576 [2024-07-15 18:09:21.109412] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:23:27.576 [2024-07-15 18:09:21.119113] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:23:27.576 [2024-07-15 18:09:21.119276] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:27.576 [2024-07-15 18:09:21.119294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x668f10 with addr=10.0.0.2, port=4420 00:23:27.576 [2024-07-15 18:09:21.119300] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x668f10 is same with the state(5) to be set 00:23:27.576 [2024-07-15 18:09:21.119310] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x668f10 (9): Bad file descriptor 00:23:27.576 [2024-07-15 18:09:21.119319] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:23:27.576 [2024-07-15 18:09:21.119324] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:23:27.576 [2024-07-15 18:09:21.119331] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:23:27.576 [2024-07-15 18:09:21.119340] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@916 -- # return 0 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@130 -- # waitforcondition '[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@912 -- # local 'cond=[[ "$(get_bdev_list)" == "nvme0n1 nvme0n2" ]]' 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@913 -- # local max=10 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@914 -- # (( max-- )) 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # eval '[[' '"$(get_bdev_list)"' == '"nvme0n1' 'nvme0n2"' ']]' 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # get_bdev_list 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:23:27.576 [2024-07-15 18:09:21.129166] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:23:27.576 [2024-07-15 18:09:21.129461] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:27.576 [2024-07-15 18:09:21.129475] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x668f10 with addr=10.0.0.2, port=4420 00:23:27.576 [2024-07-15 18:09:21.129482] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x668f10 is same with the state(5) to be set 00:23:27.576 [2024-07-15 18:09:21.129491] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x668f10 (9): Bad file descriptor 00:23:27.576 [2024-07-15 18:09:21.129506] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:23:27.576 [2024-07-15 18:09:21.129513] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:23:27.576 [2024-07-15 18:09:21.129519] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:23:27.576 [2024-07-15 18:09:21.129528] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:23:27.576 [2024-07-15 18:09:21.139217] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:23:27.576 [2024-07-15 18:09:21.139466] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:27.576 [2024-07-15 18:09:21.139479] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x668f10 with addr=10.0.0.2, port=4420 00:23:27.576 [2024-07-15 18:09:21.139486] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x668f10 is same with the state(5) to be set 00:23:27.576 [2024-07-15 18:09:21.139500] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x668f10 (9): Bad file descriptor 00:23:27.576 [2024-07-15 18:09:21.139509] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:23:27.576 [2024-07-15 18:09:21.139515] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:23:27.576 [2024-07-15 18:09:21.139522] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:23:27.576 [2024-07-15 18:09:21.139531] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:23:27.576 [2024-07-15 18:09:21.149275] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:23:27.576 [2024-07-15 18:09:21.149503] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:27.576 [2024-07-15 18:09:21.149516] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x668f10 with addr=10.0.0.2, port=4420 00:23:27.576 [2024-07-15 18:09:21.149523] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x668f10 is same with the state(5) to be set 00:23:27.576 [2024-07-15 18:09:21.149533] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x668f10 (9): Bad file descriptor 00:23:27.576 [2024-07-15 18:09:21.149544] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:23:27.576 [2024-07-15 18:09:21.149550] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:23:27.576 [2024-07-15 18:09:21.149557] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:23:27.576 [2024-07-15 18:09:21.149566] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:27.576 [2024-07-15 18:09:21.156845] bdev_nvme.c:6770:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 not found 00:23:27.576 [2024-07-15 18:09:21.156862] bdev_nvme.c:6761:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 found again 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@916 -- # return 0 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@131 -- # waitforcondition '[[ "$(get_subsystem_paths nvme0)" == "$NVMF_SECOND_PORT" ]]' 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@912 -- # local 'cond=[[ "$(get_subsystem_paths nvme0)" == "$NVMF_SECOND_PORT" ]]' 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@913 -- # local max=10 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@914 -- # (( max-- )) 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # eval '[[' '"$(get_subsystem_paths' 'nvme0)"' == '"$NVMF_SECOND_PORT"' ']]' 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # get_subsystem_paths nvme0 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@63 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers -n nvme0 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@63 -- # jq -r '.[].ctrlrs[].trid.trsvcid' 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@63 -- # sort -n 00:23:27.576 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@63 -- # xargs 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # [[ 4421 == \4\4\2\1 ]] 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@916 -- # return 0 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@132 -- # is_notification_count_eq 0 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=0 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@912 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@913 -- # local max=10 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@914 -- # (( max-- )) 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # get_notification_count 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 2 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=0 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=2 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # (( notification_count == expected_count )) 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@916 -- # return 0 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@134 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_stop_discovery -b nvme 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@136 -- # waitforcondition '[[ "$(get_subsystem_names)" == "" ]]' 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@912 -- # local 'cond=[[ "$(get_subsystem_names)" == "" ]]' 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@913 -- # local max=10 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@914 -- # (( max-- )) 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # eval '[[' '"$(get_subsystem_names)"' == '""' ']]' 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # get_subsystem_names 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_controllers 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # jq -r '.[].name' 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # xargs 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@59 -- # sort 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:27.577 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:27.835 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # [[ '' == '' ]] 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@916 -- # return 0 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@137 -- # waitforcondition '[[ "$(get_bdev_list)" == "" ]]' 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@912 -- # local 'cond=[[ "$(get_bdev_list)" == "" ]]' 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@913 -- # local max=10 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@914 -- # (( max-- )) 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # eval '[[' '"$(get_bdev_list)"' == '""' ']]' 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # get_bdev_list 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # [[ '' == '' ]] 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@916 -- # return 0 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@138 -- # is_notification_count_eq 2 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@79 -- # expected_count=2 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@80 -- # waitforcondition 'get_notification_count && ((notification_count == expected_count))' 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@912 -- # local 'cond=get_notification_count && ((notification_count == expected_count))' 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@913 -- # local max=10 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@914 -- # (( max-- )) 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # eval get_notification_count '&&' '((notification_count' == 'expected_count))' 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # get_notification_count 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@74 -- # rpc_cmd -s /tmp/host.sock notify_get_notifications -i 2 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@74 -- # jq '. | length' 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@74 -- # notification_count=2 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@75 -- # notify_id=4 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@915 -- # (( notification_count == expected_count )) 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@916 -- # return 0 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@141 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:27.836 18:09:21 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:28.771 [2024-07-15 18:09:22.472745] bdev_nvme.c:6983:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:23:28.771 [2024-07-15 18:09:22.472762] bdev_nvme.c:7063:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:23:28.771 [2024-07-15 18:09:22.472773] bdev_nvme.c:6946:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:23:29.038 [2024-07-15 18:09:22.559038] bdev_nvme.c:6912:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 new subsystem nvme0 00:23:29.297 [2024-07-15 18:09:22.824049] bdev_nvme.c:6802:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:23:29.297 [2024-07-15 18:09:22.824075] bdev_nvme.c:6761:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4421 found again 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@143 -- # NOT rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@648 -- # local es=0 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@650 -- # valid_exec_arg rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@636 -- # local arg=rpc_cmd 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@640 -- # type -t rpc_cmd 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@651 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:29.297 request: 00:23:29.297 { 00:23:29.297 "name": "nvme", 00:23:29.297 "trtype": "tcp", 00:23:29.297 "traddr": "10.0.0.2", 00:23:29.297 "adrfam": "ipv4", 00:23:29.297 "trsvcid": "8009", 00:23:29.297 "hostnqn": "nqn.2021-12.io.spdk:test", 00:23:29.297 "wait_for_attach": true, 00:23:29.297 "method": "bdev_nvme_start_discovery", 00:23:29.297 "req_id": 1 00:23:29.297 } 00:23:29.297 Got JSON-RPC error response 00:23:29.297 response: 00:23:29.297 { 00:23:29.297 "code": -17, 00:23:29.297 "message": "File exists" 00:23:29.297 } 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 1 == 0 ]] 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@651 -- # es=1 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@145 -- # get_discovery_ctrlrs 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@67 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_discovery_info 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@67 -- # jq -r '.[].name' 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@67 -- # sort 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@67 -- # xargs 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@145 -- # [[ nvme == \n\v\m\e ]] 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@146 -- # get_bdev_list 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@146 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@149 -- # NOT rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@648 -- # local es=0 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@650 -- # valid_exec_arg rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@636 -- # local arg=rpc_cmd 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@640 -- # type -t rpc_cmd 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@651 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test -w 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:29.297 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:29.297 request: 00:23:29.297 { 00:23:29.297 "name": "nvme_second", 00:23:29.297 "trtype": "tcp", 00:23:29.297 "traddr": "10.0.0.2", 00:23:29.297 "adrfam": "ipv4", 00:23:29.297 "trsvcid": "8009", 00:23:29.297 "hostnqn": "nqn.2021-12.io.spdk:test", 00:23:29.297 "wait_for_attach": true, 00:23:29.297 "method": "bdev_nvme_start_discovery", 00:23:29.297 "req_id": 1 00:23:29.297 } 00:23:29.297 Got JSON-RPC error response 00:23:29.297 response: 00:23:29.297 { 00:23:29.297 "code": -17, 00:23:29.297 "message": "File exists" 00:23:29.298 } 00:23:29.298 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 1 == 0 ]] 00:23:29.298 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@651 -- # es=1 00:23:29.298 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:23:29.298 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:23:29.298 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:23:29.298 18:09:22 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@151 -- # get_discovery_ctrlrs 00:23:29.298 18:09:22 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@67 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_discovery_info 00:23:29.298 18:09:22 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@67 -- # jq -r '.[].name' 00:23:29.298 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:29.298 18:09:22 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@67 -- # sort 00:23:29.298 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:29.298 18:09:22 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@67 -- # xargs 00:23:29.298 18:09:22 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:29.298 18:09:23 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@151 -- # [[ nvme == \n\v\m\e ]] 00:23:29.298 18:09:23 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@152 -- # get_bdev_list 00:23:29.298 18:09:23 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # sort 00:23:29.298 18:09:23 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:23:29.298 18:09:23 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # jq -r '.[].name' 00:23:29.298 18:09:23 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:29.298 18:09:23 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:29.298 18:09:23 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@55 -- # xargs 00:23:29.567 18:09:23 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:29.567 18:09:23 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@152 -- # [[ nvme0n1 nvme0n2 == \n\v\m\e\0\n\1\ \n\v\m\e\0\n\2 ]] 00:23:29.567 18:09:23 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@155 -- # NOT rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8010 -f ipv4 -q nqn.2021-12.io.spdk:test -T 3000 00:23:29.567 18:09:23 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@648 -- # local es=0 00:23:29.567 18:09:23 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@650 -- # valid_exec_arg rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8010 -f ipv4 -q nqn.2021-12.io.spdk:test -T 3000 00:23:29.567 18:09:23 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@636 -- # local arg=rpc_cmd 00:23:29.567 18:09:23 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:23:29.567 18:09:23 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@640 -- # type -t rpc_cmd 00:23:29.567 18:09:23 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:23:29.567 18:09:23 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@651 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme_second -t tcp -a 10.0.0.2 -s 8010 -f ipv4 -q nqn.2021-12.io.spdk:test -T 3000 00:23:29.567 18:09:23 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:29.567 18:09:23 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:30.544 [2024-07-15 18:09:24.064159] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:30.544 [2024-07-15 18:09:24.064186] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x682fa0 with addr=10.0.0.2, port=8010 00:23:30.544 [2024-07-15 18:09:24.064200] nvme_tcp.c:2711:nvme_tcp_ctrlr_construct: *ERROR*: failed to create admin qpair 00:23:30.544 [2024-07-15 18:09:24.064207] nvme.c: 830:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:23:30.544 [2024-07-15 18:09:24.064213] bdev_nvme.c:7045:discovery_poller: *ERROR*: Discovery[10.0.0.2:8010] could not start discovery connect 00:23:31.478 [2024-07-15 18:09:25.066557] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:23:31.478 [2024-07-15 18:09:25.066582] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x682fa0 with addr=10.0.0.2, port=8010 00:23:31.478 [2024-07-15 18:09:25.066593] nvme_tcp.c:2711:nvme_tcp_ctrlr_construct: *ERROR*: failed to create admin qpair 00:23:31.478 [2024-07-15 18:09:25.066599] nvme.c: 830:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:23:31.478 [2024-07-15 18:09:25.066605] bdev_nvme.c:7045:discovery_poller: *ERROR*: Discovery[10.0.0.2:8010] could not start discovery connect 00:23:32.413 [2024-07-15 18:09:26.068736] bdev_nvme.c:7026:discovery_poller: *ERROR*: Discovery[10.0.0.2:8010] timed out while attaching discovery ctrlr 00:23:32.413 request: 00:23:32.413 { 00:23:32.413 "name": "nvme_second", 00:23:32.413 "trtype": "tcp", 00:23:32.413 "traddr": "10.0.0.2", 00:23:32.413 "adrfam": "ipv4", 00:23:32.413 "trsvcid": "8010", 00:23:32.413 "hostnqn": "nqn.2021-12.io.spdk:test", 00:23:32.413 "wait_for_attach": false, 00:23:32.413 "attach_timeout_ms": 3000, 00:23:32.413 "method": "bdev_nvme_start_discovery", 00:23:32.413 "req_id": 1 00:23:32.413 } 00:23:32.413 Got JSON-RPC error response 00:23:32.413 response: 00:23:32.413 { 00:23:32.413 "code": -110, 00:23:32.413 "message": "Connection timed out" 00:23:32.413 } 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 1 == 0 ]] 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@651 -- # es=1 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@157 -- # get_discovery_ctrlrs 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@67 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_get_discovery_info 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@67 -- # jq -r '.[].name' 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@559 -- # xtrace_disable 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@67 -- # sort 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@67 -- # xargs 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@157 -- # [[ nvme == \n\v\m\e ]] 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@159 -- # trap - SIGINT SIGTERM EXIT 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@161 -- # kill 691568 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- host/discovery.sh@162 -- # nvmftestfini 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@488 -- # nvmfcleanup 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@117 -- # sync 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@120 -- # set +e 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@121 -- # for i in {1..20} 00:23:32.413 18:09:26 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:23:32.413 rmmod nvme_tcp 00:23:32.671 rmmod nvme_fabrics 00:23:32.671 rmmod nvme_keyring 00:23:32.671 18:09:26 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:23:32.671 18:09:26 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@124 -- # set -e 00:23:32.671 18:09:26 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@125 -- # return 0 00:23:32.671 18:09:26 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@489 -- # '[' -n 691324 ']' 00:23:32.671 18:09:26 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@490 -- # killprocess 691324 00:23:32.671 18:09:26 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@948 -- # '[' -z 691324 ']' 00:23:32.671 18:09:26 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@952 -- # kill -0 691324 00:23:32.671 18:09:26 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@953 -- # uname 00:23:32.671 18:09:26 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:23:32.671 18:09:26 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 691324 00:23:32.671 18:09:26 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:23:32.671 18:09:26 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:23:32.671 18:09:26 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@966 -- # echo 'killing process with pid 691324' 00:23:32.671 killing process with pid 691324 00:23:32.671 18:09:26 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@967 -- # kill 691324 00:23:32.671 18:09:26 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@972 -- # wait 691324 00:23:32.930 18:09:26 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:23:32.930 18:09:26 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:23:32.930 18:09:26 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:23:32.930 18:09:26 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:23:32.930 18:09:26 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@278 -- # remove_spdk_ns 00:23:32.930 18:09:26 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:32.930 18:09:26 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:23:32.930 18:09:26 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:34.833 18:09:28 nvmf_tcp.nvmf_host_discovery -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:23:34.833 00:23:34.833 real 0m17.445s 00:23:34.833 user 0m22.329s 00:23:34.833 sys 0m5.003s 00:23:34.833 18:09:28 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@1124 -- # xtrace_disable 00:23:34.833 18:09:28 nvmf_tcp.nvmf_host_discovery -- common/autotest_common.sh@10 -- # set +x 00:23:34.833 ************************************ 00:23:34.833 END TEST nvmf_host_discovery 00:23:34.833 ************************************ 00:23:34.833 18:09:28 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:23:34.833 18:09:28 nvmf_tcp -- nvmf/nvmf.sh@102 -- # run_test nvmf_host_multipath_status /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multipath_status.sh --transport=tcp 00:23:34.833 18:09:28 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:23:34.833 18:09:28 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:23:34.833 18:09:28 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:23:34.833 ************************************ 00:23:34.833 START TEST nvmf_host_multipath_status 00:23:34.833 ************************************ 00:23:34.833 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/multipath_status.sh --transport=tcp 00:23:35.091 * Looking for test storage... 00:23:35.091 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:23:35.091 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:23:35.091 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@7 -- # uname -s 00:23:35.091 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:23:35.091 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:23:35.091 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:23:35.091 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- paths/export.sh@5 -- # export PATH 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@47 -- # : 0 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@51 -- # have_pci_nics=0 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@12 -- # MALLOC_BDEV_SIZE=64 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@13 -- # MALLOC_BLOCK_SIZE=512 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@15 -- # rpc_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@16 -- # bpf_sh=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/bpftrace.sh 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@18 -- # bdevperf_rpc_sock=/var/tmp/bdevperf.sock 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@21 -- # NQN=nqn.2016-06.io.spdk:cnode1 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@31 -- # nvmftestinit 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@448 -- # prepare_net_devs 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@410 -- # local -g is_hw=no 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@412 -- # remove_spdk_ns 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@285 -- # xtrace_disable 00:23:35.092 18:09:28 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@291 -- # pci_devs=() 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@291 -- # local -a pci_devs 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@292 -- # pci_net_devs=() 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@293 -- # pci_drivers=() 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@293 -- # local -A pci_drivers 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@295 -- # net_devs=() 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@295 -- # local -ga net_devs 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@296 -- # e810=() 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@296 -- # local -ga e810 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@297 -- # x722=() 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@297 -- # local -ga x722 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@298 -- # mlx=() 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@298 -- # local -ga mlx 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:23:40.364 Found 0000:86:00.0 (0x8086 - 0x159b) 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:23:40.364 Found 0000:86:00.1 (0x8086 - 0x159b) 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@390 -- # [[ up == up ]] 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:23:40.364 Found net devices under 0000:86:00.0: cvl_0_0 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:23:40.364 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@390 -- # [[ up == up ]] 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:23:40.365 Found net devices under 0000:86:00.1: cvl_0_1 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@414 -- # is_hw=yes 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:23:40.365 18:09:33 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:23:40.365 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:23:40.365 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:23:40.365 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:23:40.365 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:23:40.624 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:23:40.624 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.234 ms 00:23:40.624 00:23:40.624 --- 10.0.0.2 ping statistics --- 00:23:40.624 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:40.624 rtt min/avg/max/mdev = 0.234/0.234/0.234/0.000 ms 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:23:40.624 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:23:40.624 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.180 ms 00:23:40.624 00:23:40.624 --- 10.0.0.1 ping statistics --- 00:23:40.624 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:23:40.624 rtt min/avg/max/mdev = 0.180/0.180/0.180/0.000 ms 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@422 -- # return 0 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@33 -- # nvmfappstart -m 0x3 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@722 -- # xtrace_disable 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@481 -- # nvmfpid=696647 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@482 -- # waitforlisten 696647 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x3 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@829 -- # '[' -z 696647 ']' 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@834 -- # local max_retries=100 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:23:40.624 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@838 -- # xtrace_disable 00:23:40.624 18:09:34 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:23:40.624 [2024-07-15 18:09:34.281747] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:23:40.624 [2024-07-15 18:09:34.281792] [ DPDK EAL parameters: nvmf -c 0x3 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:23:40.624 EAL: No free 2048 kB hugepages reported on node 1 00:23:40.624 [2024-07-15 18:09:34.339441] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:23:40.883 [2024-07-15 18:09:34.417374] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:23:40.883 [2024-07-15 18:09:34.417413] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:23:40.883 [2024-07-15 18:09:34.417422] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:23:40.883 [2024-07-15 18:09:34.417428] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:23:40.883 [2024-07-15 18:09:34.417433] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:23:40.883 [2024-07-15 18:09:34.417470] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:23:40.883 [2024-07-15 18:09:34.417473] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:23:41.451 18:09:35 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:23:41.452 18:09:35 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@862 -- # return 0 00:23:41.452 18:09:35 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:23:41.452 18:09:35 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@728 -- # xtrace_disable 00:23:41.452 18:09:35 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:23:41.452 18:09:35 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:23:41.452 18:09:35 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@34 -- # nvmfapp_pid=696647 00:23:41.452 18:09:35 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@36 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_transport -t tcp -o -u 8192 00:23:41.711 [2024-07-15 18:09:35.266222] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:23:41.711 18:09:35 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@37 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py bdev_malloc_create 64 512 -b Malloc0 00:23:41.970 Malloc0 00:23:41.970 18:09:35 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@39 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -r -m 2 00:23:41.970 18:09:35 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:23:42.229 18:09:35 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@41 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:23:42.487 [2024-07-15 18:09:35.968730] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:23:42.487 18:09:35 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@42 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 00:23:42.487 [2024-07-15 18:09:36.145240] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4421 *** 00:23:42.487 18:09:36 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@45 -- # bdevperf_pid=696971 00:23:42.487 18:09:36 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 0x4 -z -r /var/tmp/bdevperf.sock -q 128 -o 4096 -w verify -t 90 00:23:42.487 18:09:36 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@47 -- # trap 'cleanup; exit 1' SIGINT SIGTERM EXIT 00:23:42.487 18:09:36 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@48 -- # waitforlisten 696971 /var/tmp/bdevperf.sock 00:23:42.487 18:09:36 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@829 -- # '[' -z 696971 ']' 00:23:42.487 18:09:36 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bdevperf.sock 00:23:42.487 18:09:36 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@834 -- # local max_retries=100 00:23:42.487 18:09:36 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock...' 00:23:42.487 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bdevperf.sock... 00:23:42.487 18:09:36 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@838 -- # xtrace_disable 00:23:42.487 18:09:36 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:23:43.422 18:09:37 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:23:43.422 18:09:37 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@862 -- # return 0 00:23:43.422 18:09:37 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@52 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_set_options -r -1 00:23:43.681 18:09:37 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@55 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -l -1 -o 10 00:23:43.939 Nvme0n1 00:23:43.939 18:09:37 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@56 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_attach_controller -b Nvme0 -t tcp -a 10.0.0.2 -s 4421 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -x multipath -l -1 -o 10 00:23:44.514 Nvme0n1 00:23:44.514 18:09:37 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@78 -- # sleep 2 00:23:44.514 18:09:37 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@76 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -t 120 -s /var/tmp/bdevperf.sock perform_tests 00:23:46.427 18:09:39 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@90 -- # set_ANA_state optimized optimized 00:23:46.427 18:09:39 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n optimized 00:23:46.427 18:09:40 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:23:46.686 18:09:40 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@91 -- # sleep 1 00:23:47.621 18:09:41 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@92 -- # check_status true false true true true true 00:23:47.621 18:09:41 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:23:47.621 18:09:41 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:47.621 18:09:41 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:23:47.880 18:09:41 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:47.880 18:09:41 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:23:47.880 18:09:41 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:47.880 18:09:41 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:23:48.138 18:09:41 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:23:48.138 18:09:41 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:23:48.138 18:09:41 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:48.138 18:09:41 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:23:48.428 18:09:41 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:48.428 18:09:41 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:23:48.428 18:09:41 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:48.428 18:09:41 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:23:48.428 18:09:42 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:48.428 18:09:42 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:23:48.428 18:09:42 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:48.428 18:09:42 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:23:48.686 18:09:42 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:48.687 18:09:42 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:23:48.687 18:09:42 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:48.687 18:09:42 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:23:48.946 18:09:42 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:48.946 18:09:42 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@94 -- # set_ANA_state non_optimized optimized 00:23:48.946 18:09:42 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:23:48.946 18:09:42 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:23:49.204 18:09:42 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@95 -- # sleep 1 00:23:50.137 18:09:43 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@96 -- # check_status false true true true true true 00:23:50.137 18:09:43 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current false 00:23:50.137 18:09:43 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:50.137 18:09:43 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:23:50.396 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:23:50.396 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:23:50.396 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:50.396 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:23:50.655 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:50.655 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:23:50.655 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:50.655 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:23:50.913 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:50.913 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:23:50.913 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:50.913 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:23:50.913 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:50.913 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:23:50.913 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:50.913 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:23:51.172 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:51.172 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:23:51.172 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:51.172 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:23:51.431 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:51.431 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@100 -- # set_ANA_state non_optimized non_optimized 00:23:51.431 18:09:44 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:23:51.431 18:09:45 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n non_optimized 00:23:51.701 18:09:45 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@101 -- # sleep 1 00:23:52.635 18:09:46 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@102 -- # check_status true false true true true true 00:23:52.635 18:09:46 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:23:52.635 18:09:46 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:52.635 18:09:46 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:23:52.893 18:09:46 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:52.893 18:09:46 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:23:52.893 18:09:46 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:52.893 18:09:46 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:23:53.152 18:09:46 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:23:53.152 18:09:46 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:23:53.152 18:09:46 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:53.152 18:09:46 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:23:53.152 18:09:46 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:53.152 18:09:46 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:23:53.152 18:09:46 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:53.152 18:09:46 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:23:53.410 18:09:47 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:53.410 18:09:47 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:23:53.410 18:09:47 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:53.410 18:09:47 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:23:53.669 18:09:47 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:53.669 18:09:47 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:23:53.669 18:09:47 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:53.669 18:09:47 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:23:53.927 18:09:47 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:53.927 18:09:47 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@104 -- # set_ANA_state non_optimized inaccessible 00:23:53.927 18:09:47 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:23:53.927 18:09:47 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n inaccessible 00:23:54.186 18:09:47 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@105 -- # sleep 1 00:23:55.121 18:09:48 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@106 -- # check_status true false true true true false 00:23:55.121 18:09:48 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:23:55.121 18:09:48 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:55.121 18:09:48 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:23:55.379 18:09:48 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:55.379 18:09:48 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:23:55.379 18:09:48 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:55.379 18:09:48 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:23:55.637 18:09:49 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:23:55.637 18:09:49 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:23:55.637 18:09:49 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:55.637 18:09:49 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:23:55.637 18:09:49 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:55.637 18:09:49 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:23:55.637 18:09:49 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:55.638 18:09:49 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:23:55.896 18:09:49 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:55.896 18:09:49 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:23:55.896 18:09:49 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:55.896 18:09:49 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:23:56.155 18:09:49 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:56.155 18:09:49 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible false 00:23:56.155 18:09:49 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:56.155 18:09:49 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:23:56.155 18:09:49 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:23:56.155 18:09:49 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@108 -- # set_ANA_state inaccessible inaccessible 00:23:56.155 18:09:49 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n inaccessible 00:23:56.414 18:09:50 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n inaccessible 00:23:56.672 18:09:50 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@109 -- # sleep 1 00:23:57.609 18:09:51 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@110 -- # check_status false false true true false false 00:23:57.609 18:09:51 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current false 00:23:57.609 18:09:51 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:57.609 18:09:51 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:23:57.867 18:09:51 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:23:57.867 18:09:51 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:23:57.867 18:09:51 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:57.867 18:09:51 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:23:58.125 18:09:51 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:23:58.125 18:09:51 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:23:58.125 18:09:51 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:58.125 18:09:51 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:23:58.125 18:09:51 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:58.125 18:09:51 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:23:58.125 18:09:51 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:58.125 18:09:51 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:23:58.384 18:09:51 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:23:58.384 18:09:51 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible false 00:23:58.384 18:09:51 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:58.384 18:09:51 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:23:58.642 18:09:52 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:23:58.642 18:09:52 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible false 00:23:58.642 18:09:52 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:23:58.642 18:09:52 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:23:58.642 18:09:52 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:23:58.642 18:09:52 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@112 -- # set_ANA_state inaccessible optimized 00:23:58.642 18:09:52 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n inaccessible 00:23:58.901 18:09:52 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:23:59.159 18:09:52 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@113 -- # sleep 1 00:24:00.093 18:09:53 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@114 -- # check_status false true true true false true 00:24:00.093 18:09:53 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current false 00:24:00.093 18:09:53 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:00.093 18:09:53 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:24:00.352 18:09:53 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:24:00.352 18:09:53 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:24:00.352 18:09:53 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:00.352 18:09:53 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:24:00.352 18:09:54 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:00.352 18:09:54 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:24:00.352 18:09:54 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:00.352 18:09:54 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:24:00.610 18:09:54 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:00.610 18:09:54 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:24:00.610 18:09:54 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:00.610 18:09:54 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:24:00.869 18:09:54 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:00.869 18:09:54 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible false 00:24:00.869 18:09:54 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:00.869 18:09:54 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:24:01.127 18:09:54 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:24:01.127 18:09:54 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:24:01.127 18:09:54 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:01.127 18:09:54 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:24:01.127 18:09:54 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:01.127 18:09:54 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@116 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_set_multipath_policy -b Nvme0n1 -p active_active 00:24:01.386 18:09:54 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@119 -- # set_ANA_state optimized optimized 00:24:01.386 18:09:54 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n optimized 00:24:01.645 18:09:55 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:24:01.903 18:09:55 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@120 -- # sleep 1 00:24:02.838 18:09:56 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@121 -- # check_status true true true true true true 00:24:02.838 18:09:56 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:24:02.838 18:09:56 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:02.838 18:09:56 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:24:03.139 18:09:56 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:03.139 18:09:56 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:24:03.139 18:09:56 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:03.139 18:09:56 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:24:03.139 18:09:56 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:03.139 18:09:56 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:24:03.139 18:09:56 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:03.139 18:09:56 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:24:03.397 18:09:56 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:03.397 18:09:56 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:24:03.397 18:09:56 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:03.397 18:09:56 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:24:03.656 18:09:57 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:03.656 18:09:57 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:24:03.656 18:09:57 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:03.656 18:09:57 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:24:03.656 18:09:57 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:03.656 18:09:57 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:24:03.656 18:09:57 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:03.656 18:09:57 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:24:03.915 18:09:57 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:03.915 18:09:57 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@123 -- # set_ANA_state non_optimized optimized 00:24:03.915 18:09:57 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:24:04.174 18:09:57 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n optimized 00:24:04.174 18:09:57 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@124 -- # sleep 1 00:24:05.551 18:09:58 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@125 -- # check_status false true true true true true 00:24:05.552 18:09:58 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current false 00:24:05.552 18:09:58 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:05.552 18:09:58 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:24:05.552 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:24:05.552 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:24:05.552 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:05.552 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:24:05.552 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:05.552 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:24:05.552 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:05.552 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:24:05.810 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:05.810 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:24:05.810 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:05.810 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:24:06.069 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:06.069 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:24:06.069 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:06.069 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:24:06.330 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:06.330 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:24:06.330 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:06.330 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:24:06.330 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:06.330 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@129 -- # set_ANA_state non_optimized non_optimized 00:24:06.330 18:09:59 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:24:06.589 18:10:00 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n non_optimized 00:24:06.848 18:10:00 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@130 -- # sleep 1 00:24:07.784 18:10:01 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@131 -- # check_status true true true true true true 00:24:07.785 18:10:01 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:24:07.785 18:10:01 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:07.785 18:10:01 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:24:08.044 18:10:01 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:08.044 18:10:01 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current true 00:24:08.044 18:10:01 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:08.044 18:10:01 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:24:08.302 18:10:01 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:08.302 18:10:01 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:24:08.302 18:10:01 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:08.302 18:10:01 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:24:08.302 18:10:01 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:08.302 18:10:01 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:24:08.302 18:10:01 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:08.302 18:10:01 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:24:08.560 18:10:02 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:08.560 18:10:02 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:24:08.560 18:10:02 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:08.560 18:10:02 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:24:08.819 18:10:02 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:08.819 18:10:02 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible true 00:24:08.819 18:10:02 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:08.819 18:10:02 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:24:08.819 18:10:02 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:08.819 18:10:02 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@133 -- # set_ANA_state non_optimized inaccessible 00:24:08.819 18:10:02 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@59 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 -n non_optimized 00:24:09.077 18:10:02 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@60 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_subsystem_listener_set_ana_state nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4421 -n inaccessible 00:24:09.335 18:10:02 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@134 -- # sleep 1 00:24:10.270 18:10:03 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@135 -- # check_status true false true true true false 00:24:10.270 18:10:03 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@68 -- # port_status 4420 current true 00:24:10.270 18:10:03 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:10.270 18:10:03 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").current' 00:24:10.529 18:10:04 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:10.529 18:10:04 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@69 -- # port_status 4421 current false 00:24:10.529 18:10:04 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:10.529 18:10:04 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").current' 00:24:10.795 18:10:04 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:24:10.796 18:10:04 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@70 -- # port_status 4420 connected true 00:24:10.796 18:10:04 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:10.796 18:10:04 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").connected' 00:24:10.796 18:10:04 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:10.796 18:10:04 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@71 -- # port_status 4421 connected true 00:24:10.796 18:10:04 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:10.796 18:10:04 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").connected' 00:24:11.058 18:10:04 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:11.058 18:10:04 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@72 -- # port_status 4420 accessible true 00:24:11.058 18:10:04 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:11.058 18:10:04 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4420").accessible' 00:24:11.316 18:10:04 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ true == \t\r\u\e ]] 00:24:11.316 18:10:04 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@73 -- # port_status 4421 accessible false 00:24:11.316 18:10:04 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bdevperf.sock bdev_nvme_get_io_paths 00:24:11.316 18:10:04 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # jq -r '.poll_groups[].io_paths[] | select (.transport.trsvcid=="4421").accessible' 00:24:11.575 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@64 -- # [[ false == \f\a\l\s\e ]] 00:24:11.575 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@137 -- # killprocess 696971 00:24:11.575 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@948 -- # '[' -z 696971 ']' 00:24:11.575 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@952 -- # kill -0 696971 00:24:11.575 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@953 -- # uname 00:24:11.575 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:24:11.575 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 696971 00:24:11.575 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@954 -- # process_name=reactor_2 00:24:11.575 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@958 -- # '[' reactor_2 = sudo ']' 00:24:11.575 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@966 -- # echo 'killing process with pid 696971' 00:24:11.575 killing process with pid 696971 00:24:11.575 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@967 -- # kill 696971 00:24:11.575 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@972 -- # wait 696971 00:24:11.575 Connection closed with partial response: 00:24:11.575 00:24:11.575 00:24:11.861 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@139 -- # wait 696971 00:24:11.861 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@141 -- # cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:24:11.861 [2024-07-15 18:09:36.209500] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:24:11.861 [2024-07-15 18:09:36.209552] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x4 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid696971 ] 00:24:11.861 EAL: No free 2048 kB hugepages reported on node 1 00:24:11.861 [2024-07-15 18:09:36.260768] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:11.861 [2024-07-15 18:09:36.338435] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:24:11.861 Running I/O for 90 seconds... 00:24:11.861 [2024-07-15 18:09:50.046451] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:23464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.046492] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:86 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.046513] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:23472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.046521] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:9 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.046534] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:23480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.046541] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.046554] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:23488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.046560] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.046573] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:23496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.046579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:48 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.046591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:23504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.046598] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:82 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.046611] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:23512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.046617] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.046629] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:23520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.046636] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.046648] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:23528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.046655] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:53 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.046667] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:23536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.046674] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.046687] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:23544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.046699] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.046712] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:23552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.046720] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:14 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.046733] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:23560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.046740] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:125 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.046752] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:23568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.046759] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.046771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:23576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.046778] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:42 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.046790] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:23584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.046797] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:51 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.046810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:23592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.046817] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.046986] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:23600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.046995] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.047009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:23608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.047016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.047028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:23616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.047035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.047047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:23624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.861 [2024-07-15 18:09:50.047054] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:24:11.861 [2024-07-15 18:09:50.047066] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:23632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:33 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047085] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:23640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047106] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:23648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:35 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:23656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047131] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:4 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047143] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:23664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047161] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:23672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047180] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:23680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:16 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047200] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:23688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047206] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047219] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:23696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047230] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:78 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047244] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:23704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:23712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047269] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:29 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047281] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:23720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:3 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047300] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:23728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047307] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:36 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047319] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:23736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047326] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047340] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:23744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047347] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:30 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047359] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:23752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047365] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:23760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047384] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047396] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:23768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:23776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047421] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047433] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:23784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047439] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:66 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047452] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:23792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047458] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047470] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:23800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047477] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:23808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047496] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047508] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:23816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:124 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047527] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047534] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:23832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047553] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:71 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047565] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:23840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047586] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:23848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047592] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:39 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:23856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047611] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:116 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047623] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:23864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047630] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:75 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:23872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:73 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047661] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:23880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047668] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047680] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:23888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047686] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047698] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:23896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047706] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047718] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:23904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047725] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:81 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047738] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:23912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047744] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:121 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047756] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:23920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:117 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:23928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047782] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047795] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:23936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047803] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:87 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:23944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047834] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:23952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.862 [2024-07-15 18:09:50.047841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:11 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:24:11.862 [2024-07-15 18:09:50.047853] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:23960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.047860] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:12 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.047872] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:23968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.047879] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:54 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048325] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:23976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.048339] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:77 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048354] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:23984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.048361] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048373] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:23992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.048380] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048393] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:24000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.048400] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048412] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:24008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.048419] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048431] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:24016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.048438] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:60 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048450] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:24024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.048457] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:22 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048469] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:24032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.048476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:24040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.048498] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048510] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:23104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.863 [2024-07-15 18:09:50.048516] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:107 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048529] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.863 [2024-07-15 18:09:50.048535] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:58 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048547] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:23120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.863 [2024-07-15 18:09:50.048554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048566] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:23128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.863 [2024-07-15 18:09:50.048573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:24 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048585] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:23136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.863 [2024-07-15 18:09:50.048591] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:98 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048604] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:23144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.863 [2024-07-15 18:09:50.048611] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048623] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:23152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.863 [2024-07-15 18:09:50.048630] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:38 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.863 [2024-07-15 18:09:50.048650] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:56 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:23168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.863 [2024-07-15 18:09:50.048669] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:93 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048682] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:23176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.863 [2024-07-15 18:09:50.048688] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:13 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048700] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:23184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.863 [2024-07-15 18:09:50.048707] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:114 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048721] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:23192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.863 [2024-07-15 18:09:50.048728] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:103 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048740] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:23200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.863 [2024-07-15 18:09:50.048747] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048759] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:23208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.863 [2024-07-15 18:09:50.048766] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048778] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:23216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.863 [2024-07-15 18:09:50.048785] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:119 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048797] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:24048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.048803] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:96 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:24056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.048823] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048835] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:105 nsid:1 lba:24064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.048841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:105 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048853] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:24072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.048860] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:85 cdw0:0 sqhd:0040 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048872] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.048879] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048890] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:24088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.048898] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048910] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:24096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.048916] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:59 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048929] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:24104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.048935] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:2 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048947] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:24112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.048961] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:80 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048974] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:24120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.863 [2024-07-15 18:09:50.048981] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:57 cdw0:0 sqhd:0046 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.048995] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:23224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.863 [2024-07-15 18:09:50.049001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:50 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.049014] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:23232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.863 [2024-07-15 18:09:50.049021] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:120 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.049033] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:23240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.863 [2024-07-15 18:09:50.049040] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:100 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.049051] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:23248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.863 [2024-07-15 18:09:50.049058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.049070] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:23256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.863 [2024-07-15 18:09:50.049077] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:24:11.863 [2024-07-15 18:09:50.049089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:23264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049096] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:17 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:23272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049127] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:23280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049134] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049146] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:102 nsid:1 lba:23288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:102 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049165] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:72 nsid:1 lba:23296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049172] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:72 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049184] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:23304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049192] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:104 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049204] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:23312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049210] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049223] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:23320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:23328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049653] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:49 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049667] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:23336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049676] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:97 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049688] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:23344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:90 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049709] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:23352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:84 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:23360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049735] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:20 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049747] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:23368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:5 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:23376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049785] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:23384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049792] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:23392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049810] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:69 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:52 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049844] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:23408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049851] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:118 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049863] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:23416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049870] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:113 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049882] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:23424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049889] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:88 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:23432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049908] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:27 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049920] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:23440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049927] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:99 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049939] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:23448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049945] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:68 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049957] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:23456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.864 [2024-07-15 18:09:50.049964] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:23464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.864 [2024-07-15 18:09:50.049984] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:51 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.049996] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:23472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.864 [2024-07-15 18:09:50.050003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:42 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.050016] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:23480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.864 [2024-07-15 18:09:50.050023] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.050036] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:23488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.864 [2024-07-15 18:09:50.050042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:125 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.050212] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:23496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.864 [2024-07-15 18:09:50.050221] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:14 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.050245] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:23504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.864 [2024-07-15 18:09:50.050252] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.050264] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:23512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.864 [2024-07-15 18:09:50.050271] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.050283] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:23520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.864 [2024-07-15 18:09:50.050290] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:53 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.050302] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:23528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.864 [2024-07-15 18:09:50.050308] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.050320] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:23536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.864 [2024-07-15 18:09:50.050327] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.050339] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:23544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.864 [2024-07-15 18:09:50.050346] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:82 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.050358] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:23552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.864 [2024-07-15 18:09:50.050364] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:48 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.050376] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:23560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.864 [2024-07-15 18:09:50.050382] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.050395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:23568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.864 [2024-07-15 18:09:50.050402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.050414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:23576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.864 [2024-07-15 18:09:50.050420] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:9 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.050432] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:23584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.864 [2024-07-15 18:09:50.050439] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:86 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:24:11.864 [2024-07-15 18:09:50.050452] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:23592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.864 [2024-07-15 18:09:50.050460] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:54 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050474] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:23600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050480] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:12 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050493] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:23608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050500] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:11 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050512] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:23616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050519] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050531] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:23624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050537] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:87 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:23632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050557] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050569] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:23640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050575] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:117 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050587] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:23648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050594] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:121 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050606] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:23656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:81 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050625] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:23664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050631] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050643] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:23672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050650] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:23680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050669] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050681] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:23688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:73 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050699] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:23696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050708] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:75 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050720] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:23704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050726] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:116 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050738] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:23712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050745] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:39 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:23720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050766] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050779] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:23728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050785] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:71 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050798] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050806] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050819] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:23744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050825] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:124 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050838] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:23752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050856] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:23760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:23768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:23776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:66 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050912] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:23784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050919] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050931] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:23792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050939] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050951] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:23800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050958] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050970] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:23808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050977] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.050989] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:23816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.050995] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:30 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.051007] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:23824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.865 [2024-07-15 18:09:50.051014] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:24:11.865 [2024-07-15 18:09:50.051026] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:23832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.051033] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:36 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.051045] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:23840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.051051] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:3 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.051063] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:23848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.051071] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:29 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.051084] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:23856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.061914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.061931] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:23864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.061940] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:78 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.061952] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:23872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.061959] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.061971] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:23880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.061977] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:16 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.061989] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:23888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.061996] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.062009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:23896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.062016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.062029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:23904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.062035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:4 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.062047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:23912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.062053] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:35 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.062066] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:23920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.062072] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.062084] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:23928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.062091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:33 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.062103] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:23936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.062110] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.062122] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:23944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.062129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.062141] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:23952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.062147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.062159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:23960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.062166] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.062178] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:23968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.062185] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.062197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:23976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.062204] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.062216] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:23984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.062223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:104 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.062248] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:23992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.062254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:72 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.062267] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:24000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.062273] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:102 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.062886] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:24008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.062902] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.062917] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:24016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.062924] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.062937] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:24024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.062944] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:17 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.062956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:24032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.062963] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.062975] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:24040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.866 [2024-07-15 18:09:50.062982] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.062994] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:23104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.866 [2024-07-15 18:09:50.063001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:100 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.063013] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:23112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.866 [2024-07-15 18:09:50.063020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:120 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.063032] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:23120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.866 [2024-07-15 18:09:50.063039] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:50 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.063051] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.866 [2024-07-15 18:09:50.063058] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:57 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.063070] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:23136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.866 [2024-07-15 18:09:50.063076] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:80 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.063089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:23144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.866 [2024-07-15 18:09:50.063098] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:2 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.063110] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.866 [2024-07-15 18:09:50.063117] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:59 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.063129] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:23160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.866 [2024-07-15 18:09:50.063136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.063148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.866 [2024-07-15 18:09:50.063155] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.063167] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:23176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.866 [2024-07-15 18:09:50.063174] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:85 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.063186] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:23184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.866 [2024-07-15 18:09:50.063193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:105 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.063205] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:23192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.866 [2024-07-15 18:09:50.063212] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.063231] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:23200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.866 [2024-07-15 18:09:50.063238] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:96 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.063250] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:23208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.866 [2024-07-15 18:09:50.063258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:119 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:24:11.866 [2024-07-15 18:09:50.063270] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:23216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063276] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:24048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.867 [2024-07-15 18:09:50.063296] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:24056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.867 [2024-07-15 18:09:50.063315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:103 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:24064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.867 [2024-07-15 18:09:50.063336] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:114 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063348] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:24072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.867 [2024-07-15 18:09:50.063355] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:13 cdw0:0 sqhd:0040 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063367] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:24080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.867 [2024-07-15 18:09:50.063374] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:93 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063386] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:24088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.867 [2024-07-15 18:09:50.063393] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:56 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:24096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.867 [2024-07-15 18:09:50.063411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:38 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063423] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:24104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.867 [2024-07-15 18:09:50.063430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063442] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:24112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.867 [2024-07-15 18:09:50.063448] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:98 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063460] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:24120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.867 [2024-07-15 18:09:50.063467] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:24 cdw0:0 sqhd:0046 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063479] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:23224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063486] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063498] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063505] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:58 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063517] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:23240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063523] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:107 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063535] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:23248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063542] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063554] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:23256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063560] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063574] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:23264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063580] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:22 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063593] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:23272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063599] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:60 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063611] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:23280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063630] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:23288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063637] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063649] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:23296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063656] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063668] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:23304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063674] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063686] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:23312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:77 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063705] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:23320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063711] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:125 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063724] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:23328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063730] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063742] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:23336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:42 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:23344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063768] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:51 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063780] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:23352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063787] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:23360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063807] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:68 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063819] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:23368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063825] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:99 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063837] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:23376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:27 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063856] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:23384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:88 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:23392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063881] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:113 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:23400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:118 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063912] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063919] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:52 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:23416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063938] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:69 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063950] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:23424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063957] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063969] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:23432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.063988] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:23440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.063994] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:5 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.064006] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:23448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.064013] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:20 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:24:11.867 [2024-07-15 18:09:50.064025] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:23456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.867 [2024-07-15 18:09:50.064034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:84 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064046] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:90 nsid:1 lba:23464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064053] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:90 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064065] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:23472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064072] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:97 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064084] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:23480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:49 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064576] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:23488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:102 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:23496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064608] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:72 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064621] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:23504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064627] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:104 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064639] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:23512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064646] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:23520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064665] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064679] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:23528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064686] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064698] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:23536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064705] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064716] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:23544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064723] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064735] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:23552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064745] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:23560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:33 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064775] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:23568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064782] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064794] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:23576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064801] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:35 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064813] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:23584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064820] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:4 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064832] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:23592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064838] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064851] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:23600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064857] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064869] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:23608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064876] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:16 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064888] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:23616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064895] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064907] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:23624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:78 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064926] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:23632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064945] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:23640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064952] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:29 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064964] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:23648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:3 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.064986] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:23656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.064993] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:36 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.065005] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:23664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.065012] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.065024] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:23672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.065031] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:30 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.065043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:23680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.065050] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.065062] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:23688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.065068] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.065080] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:23696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.065087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.065099] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:23704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.065106] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.065118] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:23712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.065125] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:66 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.065137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:23720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.065144] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.065156] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:23728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.065163] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.065175] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:23736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.065181] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.065194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:23744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.065200] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:124 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.065214] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.065220] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.065237] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:23760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.065243] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:71 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.065256] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:23768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.065262] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.065274] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:23776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.065281] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:39 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.065294] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:23784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.868 [2024-07-15 18:09:50.065301] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:116 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:24:11.868 [2024-07-15 18:09:50.065313] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:23792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065319] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:75 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065331] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:23800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065338] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:73 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065350] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:23808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065356] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065368] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:23816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065375] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065386] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:23824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065393] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:23832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:81 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065424] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:23840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:121 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065442] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:23848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065450] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:117 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065462] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:23856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065469] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065481] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:23864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065488] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:87 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065500] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:23872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065507] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:23880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065525] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:11 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065537] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:23888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:12 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:23896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:54 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:23904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:86 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065595] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:23912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065601] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:9 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065613] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:23920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065632] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:23928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065650] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:23936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065657] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:48 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065669] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:23944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065678] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:82 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065691] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:23952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065697] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:23960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:23968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065735] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:53 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065747] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:23976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:23984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.065786] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:23992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.065793] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:14 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.066366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:24000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.066380] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:49 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.066394] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:24008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.066400] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:97 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.066413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:90 nsid:1 lba:24016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.066420] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:90 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.066432] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:24024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.066438] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:84 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.066451] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:24032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.066457] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:20 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.066470] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:24040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.869 [2024-07-15 18:09:50.066477] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:5 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.066492] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:23104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.869 [2024-07-15 18:09:50.066499] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.066511] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:23112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.869 [2024-07-15 18:09:50.066518] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.066530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:23120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.869 [2024-07-15 18:09:50.066537] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:69 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.066549] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.869 [2024-07-15 18:09:50.066555] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:52 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.066567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:23136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.869 [2024-07-15 18:09:50.066574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:118 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:24:11.869 [2024-07-15 18:09:50.066586] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:23144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.870 [2024-07-15 18:09:50.066593] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:113 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:24:11.870 [2024-07-15 18:09:50.066605] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:23152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.870 [2024-07-15 18:09:50.066612] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:88 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:24:11.870 [2024-07-15 18:09:50.066624] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:23160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.870 [2024-07-15 18:09:50.066631] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:27 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:24:11.870 [2024-07-15 18:09:50.066643] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:23168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.870 [2024-07-15 18:09:50.066649] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:99 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:24:11.870 [2024-07-15 18:09:50.066662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:23176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.870 [2024-07-15 18:09:50.066668] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:68 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:24:11.870 [2024-07-15 18:09:50.066681] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:23184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.870 [2024-07-15 18:09:50.066687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:24:11.870 [2024-07-15 18:09:50.066699] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:23192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.870 [2024-07-15 18:09:50.066706] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:51 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:24:11.870 [2024-07-15 18:09:50.066720] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:23200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.870 [2024-07-15 18:09:50.066726] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:42 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:24:11.870 [2024-07-15 18:09:50.066739] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:23208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.870 [2024-07-15 18:09:50.066745] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:24:11.870 [2024-07-15 18:09:50.066757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:23216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.870 [2024-07-15 18:09:50.066764] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:125 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:24:11.870 [2024-07-15 18:09:50.066776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:24048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.870 [2024-07-15 18:09:50.066783] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:77 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:24:11.870 [2024-07-15 18:09:50.066795] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:24056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.870 [2024-07-15 18:09:50.066802] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.066814] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:24064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.871 [2024-07-15 18:09:50.066820] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.066832] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:24072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.871 [2024-07-15 18:09:50.066839] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:0040 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.066851] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:24080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.871 [2024-07-15 18:09:50.066858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.066870] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:24088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.871 [2024-07-15 18:09:50.066877] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:60 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.066889] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:24096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.871 [2024-07-15 18:09:50.066895] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:22 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.066907] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:24104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.871 [2024-07-15 18:09:50.066914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.066928] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:24112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.871 [2024-07-15 18:09:50.066935] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.066947] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:24120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.871 [2024-07-15 18:09:50.066955] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:107 cdw0:0 sqhd:0046 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.066968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:23224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.066974] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:58 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.066986] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:23232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.066993] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.067005] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:23240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.067012] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:24 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073481] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:23248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073491] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:98 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:23256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:23264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073530] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:38 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073542] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:56 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073561] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:23280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073568] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:93 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073580] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:23288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073587] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:13 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073599] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:23296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073606] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:114 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073618] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:23304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073625] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:103 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073637] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:23312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073645] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:23320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073664] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073677] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:23328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073683] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:119 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073696] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:23336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073703] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:96 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073715] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:23344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073722] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073734] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:23352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073741] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:105 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073753] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:23360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:85 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073772] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073791] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:23376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073798] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073817] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:59 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:23392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:2 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:23400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073854] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:80 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073874] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:57 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073887] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:23416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073894] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:50 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073906] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:23424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073912] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:120 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:23432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073931] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:100 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073943] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:23440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073950] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073962] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:23448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.073981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:23456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.871 [2024-07-15 18:09:50.073987] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:17 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.074000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:23464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.871 [2024-07-15 18:09:50.074006] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.074019] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:23472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.871 [2024-07-15 18:09:50.074025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.074541] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:23480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.871 [2024-07-15 18:09:50.074554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:14 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:24:11.871 [2024-07-15 18:09:50.074569] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:23488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074576] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:23496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074596] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074608] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:23504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:53 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074630] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:23512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074651] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:23520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074658] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074670] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:23528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074677] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:82 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074688] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:23536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074695] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:48 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074707] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:23544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074714] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074726] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:23552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074744] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:23560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074751] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:9 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074763] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:23568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074770] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:86 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074782] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:23576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074789] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:54 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074801] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:23584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074808] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:12 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:23592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074827] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:11 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074839] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:23600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074846] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:23608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074866] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:87 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074878] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:23616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074885] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:23624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074904] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:117 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:23632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074923] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:121 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074935] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:23640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074941] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:81 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074954] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:23648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074960] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074972] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:23656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074979] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.074991] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:23664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.074998] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.075010] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:23672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.075016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:73 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.075029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:23680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.075035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:75 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.075048] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:23688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.075054] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:116 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.075067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:23696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.075073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:39 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.075085] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:23704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.075093] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.075106] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:23712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.075113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:71 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.075125] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.075132] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.075144] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:23728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.075151] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:124 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.075163] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:23736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.075170] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.075182] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:23744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.075189] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.075201] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:23752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.075207] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.075219] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:23760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.075231] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:66 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.075243] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:23768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.075250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.075262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:23776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.075269] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.075281] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:23784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.075288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.075300] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:23792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.075306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.075319] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:23800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.075325] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:30 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.075339] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:23808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.872 [2024-07-15 18:09:50.075345] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:24:11.872 [2024-07-15 18:09:50.075357] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:23816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075364] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:36 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075376] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:23824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075383] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:3 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:23832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:29 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:23840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075420] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075433] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:23848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075440] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:78 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075452] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:23856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075458] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075470] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:23864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075477] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:16 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:23872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075496] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075508] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:23880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075527] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:23888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075534] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:4 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:23896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075552] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:35 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075566] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:23904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075584] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:23912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075594] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:33 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075606] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:23920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075625] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:23928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075646] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:23936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075653] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075666] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:23944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075674] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075686] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:23952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075705] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:23960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075724] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:23968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075730] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:104 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075743] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:23976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:72 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.075761] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:23984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.075768] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:102 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.076361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:23992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.076375] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.076390] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:24000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.076399] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.076412] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:24008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.076418] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:17 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.076431] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:24016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.076437] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.076450] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:24024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.076456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.076468] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:100 nsid:1 lba:24032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.076475] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:100 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.076487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:120 nsid:1 lba:24040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.873 [2024-07-15 18:09:50.076494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:120 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.076506] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:23104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.873 [2024-07-15 18:09:50.076513] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:50 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.076526] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.873 [2024-07-15 18:09:50.076533] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:57 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.076545] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:23120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.873 [2024-07-15 18:09:50.076552] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:80 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.076565] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:23128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.873 [2024-07-15 18:09:50.076571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:2 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.076584] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.873 [2024-07-15 18:09:50.076590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:59 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.076603] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:23144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.873 [2024-07-15 18:09:50.076609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.076622] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.873 [2024-07-15 18:09:50.076631] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.076643] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:23160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.873 [2024-07-15 18:09:50.076650] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:85 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.076662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:23168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.873 [2024-07-15 18:09:50.076668] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:105 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.076680] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:23176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.873 [2024-07-15 18:09:50.076687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.076699] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:23184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.873 [2024-07-15 18:09:50.076706] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:96 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.076718] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:23192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.873 [2024-07-15 18:09:50.076724] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:119 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:24:11.873 [2024-07-15 18:09:50.076737] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:23200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.076743] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.076755] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:23208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.076762] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.076774] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:23216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.076781] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:103 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.076793] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:24048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.874 [2024-07-15 18:09:50.076799] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:114 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.076811] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:24056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.874 [2024-07-15 18:09:50.076818] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:13 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.076830] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:24064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.874 [2024-07-15 18:09:50.076837] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:93 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.076850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:24072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.874 [2024-07-15 18:09:50.076860] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:56 cdw0:0 sqhd:0040 p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.076875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:24080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.874 [2024-07-15 18:09:50.076883] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:38 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.076899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:24088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.874 [2024-07-15 18:09:50.076907] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.076921] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:24096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.874 [2024-07-15 18:09:50.076929] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:98 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.076944] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:24104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.874 [2024-07-15 18:09:50.076953] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:24 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.076966] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:24112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.874 [2024-07-15 18:09:50.076973] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.076985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:24120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.874 [2024-07-15 18:09:50.076993] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:58 cdw0:0 sqhd:0046 p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077005] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:23224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077012] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:107 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077024] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:23232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077031] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077044] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:23240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077051] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077063] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:23248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077069] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:22 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077081] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:23256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077088] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:60 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077101] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:23264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077107] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077121] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:23272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077128] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077140] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:23280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077147] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:23288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077166] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077178] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:23296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077184] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:77 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:23304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077204] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:125 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077216] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:23312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077241] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:23320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077247] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:42 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077260] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:23328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077266] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:51 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077279] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:23336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077285] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077297] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:23344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077304] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:68 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077316] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:23352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077323] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:99 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077335] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:23360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077342] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:27 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077358] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:23368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077365] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:88 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077379] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:23376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077386] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:113 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077398] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:23384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077405] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:118 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077417] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077424] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:52 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:24:11.874 [2024-07-15 18:09:50.077436] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:23400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.874 [2024-07-15 18:09:50.077443] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:69 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.077455] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:23408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.875 [2024-07-15 18:09:50.077463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.077475] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:23416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.875 [2024-07-15 18:09:50.077483] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.077495] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:23424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.875 [2024-07-15 18:09:50.077502] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:5 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.077516] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:23432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.875 [2024-07-15 18:09:50.077522] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:20 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.077535] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:23440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.875 [2024-07-15 18:09:50.077541] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:84 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.077555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:23448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.875 [2024-07-15 18:09:50.077562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:90 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.077574] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:23456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.875 [2024-07-15 18:09:50.077581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:97 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.077596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:23464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.077604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:49 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:23472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:102 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078106] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:23480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:72 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078125] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:23488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078132] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:104 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078144] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:23496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078151] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078163] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:23504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078171] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078185] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:23512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078192] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078204] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:23520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078212] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078230] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:23528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078237] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078249] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:23536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078269] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:23544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078275] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:33 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078287] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:23552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078295] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078307] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:23560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:35 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078328] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:23568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:4 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078347] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:23576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078354] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:23584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078374] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078386] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:23592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078393] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:16 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:23600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078424] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:23608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:78 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078443] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:23616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078450] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078462] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:23624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078469] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:29 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078480] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:23632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078487] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:3 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078500] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:23640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078507] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:36 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:23648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078526] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:23656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:30 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078560] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:23664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078566] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078579] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:23672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078585] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078597] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:23680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078616] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:23688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078623] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078635] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:23696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078642] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:66 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078654] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:23704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078661] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078673] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:23712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078680] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078692] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:23720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078699] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:23728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.875 [2024-07-15 18:09:50.078718] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:124 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:24:11.875 [2024-07-15 18:09:50.078730] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.078737] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.078749] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:23744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.078757] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:71 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.078768] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:23752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.078775] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.078788] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:23760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.078795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:39 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.078808] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:23768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.078814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:116 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.078826] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:23776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.078835] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:75 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.078847] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:23784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.078854] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:73 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.078866] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:23792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.078872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.078884] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:23800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.078891] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.078903] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:23808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.078910] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.078922] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:23816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.078928] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:81 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.078940] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:23824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.078947] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:121 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.078959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:23832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.078966] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:117 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.078978] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:23840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.078985] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.078997] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:23848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079004] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:87 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079016] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:23856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079024] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079036] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:23864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079043] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:11 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079055] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:23872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079063] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:12 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079075] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:23880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:54 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079094] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:23888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079100] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:86 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079113] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:23896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079119] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:9 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079132] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:23904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079140] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079152] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:23912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079170] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:23920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:48 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079190] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:23928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:82 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079208] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:23936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079215] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079232] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:23944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079239] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:23952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:53 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079271] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:23960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079278] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079290] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:23968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079309] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:23976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:14 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079893] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:23984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079906] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:49 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079921] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:23992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079928] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:97 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079940] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:90 nsid:1 lba:24000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079947] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:90 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079960] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:24008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:84 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079979] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:24016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.079986] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:20 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.079998] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:24024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.080005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:5 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.080017] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:24032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.080024] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.080036] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:24040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.876 [2024-07-15 18:09:50.080043] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.080055] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:23104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.876 [2024-07-15 18:09:50.080062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:69 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.080077] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.876 [2024-07-15 18:09:50.080084] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:52 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:24:11.876 [2024-07-15 18:09:50.080096] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:23120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.876 [2024-07-15 18:09:50.080103] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:118 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080115] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:23128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080122] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:113 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080134] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:23136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080140] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:88 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080152] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:23144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080159] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:27 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:23152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080180] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:99 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080192] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:23160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080199] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:68 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080211] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:23168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080218] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080235] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:23176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:51 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080254] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:23184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080261] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:42 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080274] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:23192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080281] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080293] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:23200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080300] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:125 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080313] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:23208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080320] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:77 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080332] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:23216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080340] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080352] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:24048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.877 [2024-07-15 18:09:50.080359] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080371] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:24056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.877 [2024-07-15 18:09:50.080378] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080391] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:24064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.877 [2024-07-15 18:09:50.080397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080409] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:24072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.877 [2024-07-15 18:09:50.080416] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:60 cdw0:0 sqhd:0040 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080428] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:24080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.877 [2024-07-15 18:09:50.080435] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:22 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080447] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:24088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.877 [2024-07-15 18:09:50.080454] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080466] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:24096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.877 [2024-07-15 18:09:50.080473] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080485] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:24104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.877 [2024-07-15 18:09:50.080493] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:107 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080505] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:24112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.877 [2024-07-15 18:09:50.080512] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:58 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080524] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:24120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.877 [2024-07-15 18:09:50.080531] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0046 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080543] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:23224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080551] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:24 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080564] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:23232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:98 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080583] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:23240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080590] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080602] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:23248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080609] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:38 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080621] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080628] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:56 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:23264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080648] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:93 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080660] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:23272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080667] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:13 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080679] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:23280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080686] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:114 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080698] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:23288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:103 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:23296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080723] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080736] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:23304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080742] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080754] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:23312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080761] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:119 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080773] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:23320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080782] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:96 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080794] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:23328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080802] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:23336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:105 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080834] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:23344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:85 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080853] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080860] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.080872] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:23360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.080879] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:24:11.877 [2024-07-15 18:09:50.084847] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.877 [2024-07-15 18:09:50.084858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:59 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.084872] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:23376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.878 [2024-07-15 18:09:50.084878] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:2 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.084891] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:23384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.878 [2024-07-15 18:09:50.084897] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:80 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.084910] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.878 [2024-07-15 18:09:50.084917] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:57 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.084929] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:23400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.878 [2024-07-15 18:09:50.084936] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:50 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.084948] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:23408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.878 [2024-07-15 18:09:50.084955] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:120 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.084967] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:23416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.878 [2024-07-15 18:09:50.084974] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:100 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.084987] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:23424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.878 [2024-07-15 18:09:50.084994] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085007] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:23432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.878 [2024-07-15 18:09:50.085013] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085026] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:23440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.878 [2024-07-15 18:09:50.085033] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:17 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085045] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:23448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.878 [2024-07-15 18:09:50.085052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085064] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:23456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.878 [2024-07-15 18:09:50.085072] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085581] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:23464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085595] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:14 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085609] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:23472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085616] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085628] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:23480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085635] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085648] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:23488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085655] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:53 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085667] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:23496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085674] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085687] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:23504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085706] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:23512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085713] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:82 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:23520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085735] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:48 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085746] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:23528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085753] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:23536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085785] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:23544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085792] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:9 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:23552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085812] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:86 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085824] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:23560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:54 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085843] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:23568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085849] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:12 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085861] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:23576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085868] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:11 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085881] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:23584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085887] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:23592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085906] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:87 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085919] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:23600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085925] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085938] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:23608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085944] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:117 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085957] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:23616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085965] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:121 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085977] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:23624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.085984] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:81 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.085996] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:23632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.086003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.086015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:23640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.086022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.086035] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:23648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.086041] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.086053] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:23656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.086060] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:73 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.086072] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:23664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.086079] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:75 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.086091] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:23672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.086098] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:116 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.086110] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:23680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.086117] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:39 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.086129] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:23688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.086136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:11.878 [2024-07-15 18:09:50.086148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:23696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.878 [2024-07-15 18:09:50.086155] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:71 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086168] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086174] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086187] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:23712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086195] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:124 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086207] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:23720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086214] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086231] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:23728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086238] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086250] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:23736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086257] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086269] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:23744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086276] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:66 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086288] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:23752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086295] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086307] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:23760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086314] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086326] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:23768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086333] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086345] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:23776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086352] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086364] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:23784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:30 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086383] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:23792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086402] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:23800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086409] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:36 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086421] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:23808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:3 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086443] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:23816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086450] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:29 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086463] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:23824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086469] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086482] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:23832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086489] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:78 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086501] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:23840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086520] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:23848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086527] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:16 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:23856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086558] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:23864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086565] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:23872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086583] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:4 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:23880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086602] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:35 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:23888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086622] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:23896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086641] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:33 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086653] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:23904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086660] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086674] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:23912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086680] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086693] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:23920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086699] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:23928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086718] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086730] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:23936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086737] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086749] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:23944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086768] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:23952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086775] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:104 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086787] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:23960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:72 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.086807] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:23968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.086814] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:102 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.087379] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:23976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.879 [2024-07-15 18:09:50.087391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:24:11.879 [2024-07-15 18:09:50.087405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:23984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.880 [2024-07-15 18:09:50.087412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087425] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:23992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.880 [2024-07-15 18:09:50.087432] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:17 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087444] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:24000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.880 [2024-07-15 18:09:50.087451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087463] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:24008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.880 [2024-07-15 18:09:50.087472] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087484] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:100 nsid:1 lba:24016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.880 [2024-07-15 18:09:50.087491] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:100 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087504] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:120 nsid:1 lba:24024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.880 [2024-07-15 18:09:50.087510] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:120 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:50 nsid:1 lba:24032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.880 [2024-07-15 18:09:50.087529] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:50 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087542] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:24040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.880 [2024-07-15 18:09:50.087549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:57 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087561] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:23104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.087568] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:80 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087580] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:23112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.087587] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:2 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087599] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.087606] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:59 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087618] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:23128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.087625] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087637] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.087644] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087657] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:23144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.087663] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:85 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087676] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:23152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.087683] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:105 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087695] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:23160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.087703] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087715] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:23168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.087722] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:96 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087734] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:23176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.087741] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:119 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087753] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:23184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.087760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087772] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:23192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.087779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087791] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:23200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.087798] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:103 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:23208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.087816] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:114 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087828] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:23216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.087835] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:13 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087847] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:24048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.880 [2024-07-15 18:09:50.087854] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:93 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087866] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:24056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.880 [2024-07-15 18:09:50.087873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:56 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087885] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:24064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.880 [2024-07-15 18:09:50.087892] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:38 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087904] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:24072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.880 [2024-07-15 18:09:50.087911] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0040 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:24080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.880 [2024-07-15 18:09:50.087930] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:98 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087943] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:24088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.880 [2024-07-15 18:09:50.087950] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:24 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087962] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:24096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.880 [2024-07-15 18:09:50.087969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.087981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:24104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.880 [2024-07-15 18:09:50.087988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:58 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.088000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:24112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.880 [2024-07-15 18:09:50.088007] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:107 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.088019] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:24120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.880 [2024-07-15 18:09:50.088026] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:0046 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.088038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:23224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.088045] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.088057] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:23232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.088064] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:22 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.088076] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:23240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.088083] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:60 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.088095] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:23248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.088102] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.088114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:23256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.088120] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.088133] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:23264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.088139] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.088152] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:23272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.088159] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.088172] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:23280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.088179] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:77 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:24:11.880 [2024-07-15 18:09:50.088191] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:23288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.880 [2024-07-15 18:09:50.088198] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:125 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.088210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:23296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.088216] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.088232] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:23304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.088239] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:42 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.088252] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:23312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.088258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:51 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.088271] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:23320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.088277] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.088289] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:23328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.088296] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:68 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.088308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:23336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.088315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:99 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.088327] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:23344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.088334] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:27 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.088347] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:23352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.088353] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:88 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.088365] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:23360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.088372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:113 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.088385] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:23368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.088391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:118 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.088405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.088412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:52 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.088424] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:23384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.088431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:69 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.088444] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:23392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.088451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.088463] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:23400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.088469] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.088481] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:23408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.088488] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:5 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.088500] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:23416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.088507] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:20 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.088519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:23424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.088525] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:84 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.088538] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:23432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.088544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:90 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.088556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:23440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.088563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:97 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.088575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:23448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.088581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:49 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089059] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:102 nsid:1 lba:23456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.881 [2024-07-15 18:09:50.089069] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:102 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:23464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.881 [2024-07-15 18:09:50.089089] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:72 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089101] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:23472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.881 [2024-07-15 18:09:50.089110] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:104 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089122] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:23480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.881 [2024-07-15 18:09:50.089129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089142] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:23488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.881 [2024-07-15 18:09:50.089149] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089161] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:23496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.881 [2024-07-15 18:09:50.089167] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089179] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:23504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.881 [2024-07-15 18:09:50.089186] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089199] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:23512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.881 [2024-07-15 18:09:50.089206] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089218] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:23520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.881 [2024-07-15 18:09:50.089229] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089242] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:23528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.881 [2024-07-15 18:09:50.089250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:33 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:23536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.881 [2024-07-15 18:09:50.089269] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089281] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:23544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.881 [2024-07-15 18:09:50.089288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:35 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089300] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:23552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.881 [2024-07-15 18:09:50.089306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:4 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089318] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:23560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.881 [2024-07-15 18:09:50.089325] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:23568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.881 [2024-07-15 18:09:50.089346] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089359] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:23576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.881 [2024-07-15 18:09:50.089365] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:16 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089378] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:23584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.881 [2024-07-15 18:09:50.089385] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089397] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:23592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.881 [2024-07-15 18:09:50.089403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:78 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089416] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:23600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.881 [2024-07-15 18:09:50.089422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:23608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.881 [2024-07-15 18:09:50.089441] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:29 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:24:11.881 [2024-07-15 18:09:50.089453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:23616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.881 [2024-07-15 18:09:50.089460] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:3 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089472] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:23624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089478] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:36 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089492] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:23632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089499] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089511] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:23640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:30 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:23648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089536] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089549] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:23656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089555] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:23664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089587] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:23672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089594] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089606] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:23680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089613] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:66 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089625] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:23688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089644] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:23696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089651] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089663] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:23704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089670] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089682] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:23712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089689] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:124 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089701] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089708] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089720] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:23728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089727] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:71 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089739] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:23736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089746] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089758] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:23744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089765] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:39 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089777] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:23752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:116 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089798] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:23760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089804] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:75 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089818] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:23768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089825] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:73 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089837] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:23776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089844] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089856] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:23784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:23792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:23800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:81 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089913] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:23808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089920] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:121 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:23816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089938] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:117 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089951] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:23824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089958] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089970] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:23832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089976] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:87 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.089989] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:23840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.089996] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.090008] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:23848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.090015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:11 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.090028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:23856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.090034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:12 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.090047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:23864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.090055] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:54 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.090067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:23872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.090074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:86 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.090087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:23880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.090093] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:9 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.090107] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:23888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.090114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.090126] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:95 nsid:1 lba:23896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.090133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.090639] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:23904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.090652] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:48 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.090666] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:23912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.090674] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:82 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.090686] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:23920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.090693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.090706] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:23928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.090712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.090725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:23936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.090731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:53 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.090744] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:23944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.882 [2024-07-15 18:09:50.090751] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:11.882 [2024-07-15 18:09:50.090763] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:23952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.090770] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.090782] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:23960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.090792] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:14 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.090804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:23968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.090811] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:49 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.090823] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:23976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.090830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:97 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.090842] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:90 nsid:1 lba:23984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.090850] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:90 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.090862] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:23992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.090869] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:84 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.090881] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:24000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.090888] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:20 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.090900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:24008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.090907] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:5 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.090920] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:24016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.090927] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.090939] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:24024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.090946] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.090958] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:24032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.090964] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:69 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.090977] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:24040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.090983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:52 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.090996] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:23104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.883 [2024-07-15 18:09:50.091003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:118 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:23112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.883 [2024-07-15 18:09:50.091022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:113 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091035] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:23120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.883 [2024-07-15 18:09:50.091043] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:88 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091055] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:23128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.883 [2024-07-15 18:09:50.091062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:27 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091074] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:23136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.883 [2024-07-15 18:09:50.091081] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:99 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091093] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:23144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.883 [2024-07-15 18:09:50.091101] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:68 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:23152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.883 [2024-07-15 18:09:50.091121] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091133] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:23160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.883 [2024-07-15 18:09:50.091140] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:51 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091152] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:23168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.883 [2024-07-15 18:09:50.091159] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:42 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:23176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.883 [2024-07-15 18:09:50.091178] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091190] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:23184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.883 [2024-07-15 18:09:50.091197] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:125 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091209] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:23192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.883 [2024-07-15 18:09:50.091216] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:77 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091233] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:23200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.883 [2024-07-15 18:09:50.091240] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091252] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:23208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.883 [2024-07-15 18:09:50.091259] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091273] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:23216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.883 [2024-07-15 18:09:50.091280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:24048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.091299] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091311] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:24056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.091318] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:60 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091330] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:24064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.091337] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:22 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091349] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:24072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.091356] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:0040 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091368] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:24080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.091375] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091687] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:24088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.091697] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:107 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:24096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.091718] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:58 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091730] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:24104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.091737] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091749] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:24112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.091756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:24 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:24:11.883 [2024-07-15 18:09:50.091768] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:24120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.883 [2024-07-15 18:09:50.091775] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:98 cdw0:0 sqhd:0046 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.091787] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:23224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.091793] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.091805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:23232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.091817] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:38 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.091829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:23240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.091836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:56 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.091848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:23248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.091855] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:93 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.091867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:23256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.091874] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:13 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.091886] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:23264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.091893] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:114 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.091905] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:23272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.091912] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:103 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.091924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:23280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.091931] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.091943] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:23288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.091949] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.091961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:23296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.091968] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:119 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.091980] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:23304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.091987] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:96 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.091999] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:23312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.092006] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092018] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:23320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.092025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:105 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092037] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:23328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.092045] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:85 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092057] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.092064] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092076] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:23344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.092082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092095] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.092101] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:59 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:23360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.092120] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:2 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092132] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:23368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.092139] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:80 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092151] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.092158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:57 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:23384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.092177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:50 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092190] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:23392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.092196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:120 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092208] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:23400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.092215] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:100 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092232] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:23408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.092239] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092251] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:23416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.092258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092270] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:23424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.092277] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:17 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092290] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:23432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.092297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092310] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:23440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.092316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092328] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:23448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.092335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092347] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:23456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.884 [2024-07-15 18:09:50.092355] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092367] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:23464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.884 [2024-07-15 18:09:50.092374] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:9 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092386] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:23472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.884 [2024-07-15 18:09:50.092393] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:86 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:54 nsid:1 lba:23480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.884 [2024-07-15 18:09:50.092412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:54 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092424] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:23488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.884 [2024-07-15 18:09:50.092431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:12 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092442] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:23496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.884 [2024-07-15 18:09:50.092449] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:11 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092462] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:23504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.884 [2024-07-15 18:09:50.092469] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092481] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:23512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.884 [2024-07-15 18:09:50.092487] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:87 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092500] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:23520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.884 [2024-07-15 18:09:50.092506] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092520] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:23528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.884 [2024-07-15 18:09:50.092527] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:117 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:23536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.884 [2024-07-15 18:09:50.092546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:121 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092558] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:23544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.884 [2024-07-15 18:09:50.092564] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:81 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:24:11.884 [2024-07-15 18:09:50.092577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:23552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.092583] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.092595] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:23560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.092602] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.092615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:23568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.092621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.092634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:23576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.092640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:73 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.092652] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:23584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.092659] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:75 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.092671] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:23592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.092678] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:116 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.092690] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:23600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.092697] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:39 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.092709] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:23608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.092716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.092728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:23616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.092734] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:71 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.092746] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.092754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.092766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:23632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.092773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:124 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.092786] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:23640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.092792] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093258] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:23648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093269] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093283] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:23656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093290] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093302] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:23664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093309] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:66 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093321] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:23672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093328] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093341] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:23680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093347] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093360] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:23688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093367] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093378] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:23696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093385] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093397] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:23704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093406] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:30 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093418] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:23712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093425] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093437] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:23720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093447] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:36 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093460] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:23728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093467] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:3 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093479] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:23736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093486] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:29 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093498] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:23744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093505] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093517] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:23752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093524] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:78 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093536] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:23760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093543] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093555] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:23768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:16 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093574] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:23776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093593] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:23784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093612] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:23792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:4 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093633] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:23800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:35 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093652] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:23808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093659] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093672] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:23816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093679] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:33 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093692] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:23824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093699] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093712] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:23832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093719] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093731] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:23840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093738] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093751] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:23848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093757] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093770] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:23856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:23864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093796] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093808] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:23872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093815] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:104 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:24:11.885 [2024-07-15 18:09:50.093827] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:23880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.885 [2024-07-15 18:09:50.093834] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:72 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.093846] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:23888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.093853] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:102 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.093865] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:23896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.093872] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.093885] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:23904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.093892] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.093904] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:23912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.093910] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:22 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.093924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:23920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.093931] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:60 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.093943] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:23928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.093950] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.093965] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:23936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.093971] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.093984] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:23944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.093990] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094003] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:23952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.094010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094022] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:23960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.094029] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:77 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094041] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:23968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.094048] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:125 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094060] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:23976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.094067] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094079] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:23984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.094086] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:42 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094098] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:23992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.094105] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:51 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:24000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.094124] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:24008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.094143] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:68 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:24016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.094163] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:99 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094176] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:24024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.094183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:27 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094629] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:24032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.094642] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:88 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094656] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:113 nsid:1 lba:24040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.094663] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:113 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094676] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:23104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.886 [2024-07-15 18:09:50.094687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:118 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094699] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.886 [2024-07-15 18:09:50.094706] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:52 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094719] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:23120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.886 [2024-07-15 18:09:50.094726] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:69 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094738] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:23128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.886 [2024-07-15 18:09:50.094745] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:23136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.886 [2024-07-15 18:09:50.094764] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:23144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.886 [2024-07-15 18:09:50.094784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:5 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094796] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:23152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.886 [2024-07-15 18:09:50.094803] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:20 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:23160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.886 [2024-07-15 18:09:50.094822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:84 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094834] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:23168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.886 [2024-07-15 18:09:50.094843] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:90 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094855] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:23176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.886 [2024-07-15 18:09:50.094862] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:97 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094874] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:23184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.886 [2024-07-15 18:09:50.094881] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:49 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094893] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:23192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.886 [2024-07-15 18:09:50.094900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:14 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094912] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:23200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.886 [2024-07-15 18:09:50.094919] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094931] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:23208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.886 [2024-07-15 18:09:50.094938] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094951] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.886 [2024-07-15 18:09:50.094957] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:53 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094970] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:24048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.094976] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.094988] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:24056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.094997] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.095009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:24064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.095016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:82 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.095028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:24072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.095035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:48 cdw0:0 sqhd:0040 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.095047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:24080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.095053] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:11.886 [2024-07-15 18:09:50.095065] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:24088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.886 [2024-07-15 18:09:50.095072] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:124 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.095086] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.887 [2024-07-15 18:09:50.095093] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.095106] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:24104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.887 [2024-07-15 18:09:50.095112] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:71 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.095125] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:24112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.887 [2024-07-15 18:09:50.095131] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.095144] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:24120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.887 [2024-07-15 18:09:50.095150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:39 cdw0:0 sqhd:0046 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.098958] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:23224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.098969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:116 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.098982] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:23232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.098989] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:75 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:23240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:73 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099023] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:23248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099030] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099042] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:23256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099049] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099061] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:23264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099068] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099080] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:81 nsid:1 lba:23272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:81 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099099] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:23280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099107] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:121 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099120] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:23288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099127] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:117 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099140] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:23296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099146] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099158] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:23304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099165] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:87 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099178] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:23312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099184] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:23320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099204] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:11 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099216] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:23328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:12 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099240] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099247] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:54 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099260] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:23344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:86 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099279] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:23352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099286] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:9 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099299] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:23360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099307] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099319] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:23368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099326] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099338] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:23376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099345] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099357] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:23384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099366] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099379] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:23392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099385] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:17 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099398] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:23400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099417] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:23408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099424] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099436] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:23416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099443] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:100 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099455] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:23424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099462] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:120 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099474] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:23432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:50 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099493] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099500] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:57 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099512] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:23448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099519] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:80 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099531] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:23456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.887 [2024-07-15 18:09:50.099538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:2 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:24:11.887 [2024-07-15 18:09:50.099551] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:23464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.887 [2024-07-15 18:09:50.099557] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:59 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:23472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100026] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100040] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100050] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100063] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:23488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100070] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:85 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:105 nsid:1 lba:23496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100089] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:105 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100101] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:23504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100120] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:23512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100127] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:96 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100140] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:119 nsid:1 lba:23520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100146] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:119 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:23528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100166] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100178] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:23536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100185] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:23544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100204] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:103 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100216] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:23552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:114 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100241] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:23560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100248] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:13 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100260] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:23568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:93 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100279] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:23576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100286] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:56 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100300] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:23584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100307] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:38 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100319] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:23592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100326] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100338] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:23600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100345] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:98 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100358] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:23608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100365] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:24 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:23616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100384] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100396] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:23624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:58 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100415] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:23632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:107 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:23640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100441] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:27 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:23648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100460] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:99 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100472] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:23656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:68 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100491] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:23664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100498] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100511] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:23672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:51 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100533] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:23680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100540] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:42 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100552] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:23688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100559] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100571] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:23696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100578] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:125 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:23704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100597] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:77 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100610] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:23712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100616] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100629] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:23720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100636] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100647] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:23728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100654] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100667] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:23736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100673] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100686] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:23744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:60 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100705] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:23752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:22 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100724] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:23760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100743] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:23768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100750] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:23776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100770] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:102 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100782] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:23784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100789] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:72 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100801] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:23792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.888 [2024-07-15 18:09:50.100809] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:104 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:24:11.888 [2024-07-15 18:09:50.100821] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:23800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.100828] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.100841] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:23808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.100847] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.100860] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:23816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.100867] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.100879] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:23824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.100886] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.100899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:23832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.100906] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.100918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:23840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.100925] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.100937] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:23848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.100944] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:33 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.100956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:23856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.100963] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.100975] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:23864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.100982] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:35 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.100994] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:23872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101002] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:4 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.101015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:23880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.101034] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:23888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101040] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.101052] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:23896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101060] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:16 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.101072] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:23904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101079] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.101091] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:23912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101098] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:78 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.101109] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:23920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101117] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.101129] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:23928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101135] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:29 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.101148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:23936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:3 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.101167] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:23944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101173] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:36 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.101185] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:23952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101192] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.101205] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:23960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101211] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:30 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.101227] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:23968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.101248] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:23976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101255] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.101268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:23984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101275] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.101287] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:23992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101294] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.101307] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:24000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101314] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:66 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.101326] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:24008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101333] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.101346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:24016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101353] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.101964] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:59 nsid:1 lba:24024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101977] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:59 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.101991] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:24032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.101998] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:2 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.102010] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:24040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.889 [2024-07-15 18:09:50.102017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:80 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.102030] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.889 [2024-07-15 18:09:50.102037] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:57 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.102049] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:23112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.889 [2024-07-15 18:09:50.102056] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:50 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.102068] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:23120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.889 [2024-07-15 18:09:50.102075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:120 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.102089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:23128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.889 [2024-07-15 18:09:50.102096] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:100 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.102108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:23136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.889 [2024-07-15 18:09:50.102115] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.102127] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:23144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.889 [2024-07-15 18:09:50.102134] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.102147] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:23152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.889 [2024-07-15 18:09:50.102154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:17 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.102167] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:23160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.889 [2024-07-15 18:09:50.102174] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.102186] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:23168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.889 [2024-07-15 18:09:50.102193] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.102205] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:23176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.889 [2024-07-15 18:09:50.102212] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:24:11.889 [2024-07-15 18:09:50.102229] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:23184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.889 [2024-07-15 18:09:50.102236] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102249] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:23192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102255] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:9 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102268] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:23200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102274] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:86 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102293] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:54 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102305] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:23216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102312] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:12 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102324] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:24048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.890 [2024-07-15 18:09:50.102333] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:11 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102345] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:112 nsid:1 lba:24056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.890 [2024-07-15 18:09:50.102352] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102364] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:24064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.890 [2024-07-15 18:09:50.102371] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:87 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:24072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.890 [2024-07-15 18:09:50.102390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0040 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:117 nsid:1 lba:24080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.890 [2024-07-15 18:09:50.102409] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:117 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102421] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:121 nsid:1 lba:24088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.890 [2024-07-15 18:09:50.102428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:121 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102440] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:24096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.890 [2024-07-15 18:09:50.102447] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:81 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102459] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:7 nsid:1 lba:24104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.890 [2024-07-15 18:09:50.102466] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102478] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:24112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.890 [2024-07-15 18:09:50.102485] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102498] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:24120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.890 [2024-07-15 18:09:50.102505] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:0046 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102517] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:23224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102524] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:73 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102536] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:23232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102543] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:75 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:23240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102564] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:116 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102577] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:23248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102584] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:39 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:23256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102602] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:23264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:71 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102633] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:23272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102652] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:124 nsid:1 lba:23280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102660] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:124 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102672] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:23288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102679] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102691] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:23296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102698] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:48 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:82 nsid:1 lba:23304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102717] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:82 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102730] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:23312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102737] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102750] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:23320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102775] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:53 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102788] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:23336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102796] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102809] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:23344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102817] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:23352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102837] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:14 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102849] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:23360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102856] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:49 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102868] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:23368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102875] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:97 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102888] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:23376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102895] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:90 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102908] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:23384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102915] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:84 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102927] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:23392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102934] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:20 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102946] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:23400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102954] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:5 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102966] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:23408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102973] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.102985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:23416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.102992] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.103004] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:23424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.103011] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:69 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.103023] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.890 [2024-07-15 18:09:50.103029] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:52 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:11.890 [2024-07-15 18:09:50.103043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:23440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.891 [2024-07-15 18:09:50.103050] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:118 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.103063] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:23448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.891 [2024-07-15 18:09:50.103070] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:113 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.103082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:23456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.891 [2024-07-15 18:09:50.103089] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:88 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.103528] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:23464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.103538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.103551] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:23472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.103558] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.103571] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:23480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.103578] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:66 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.103590] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:23488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.103596] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.103609] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:23496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.103615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.103628] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:23504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.103634] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.103646] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:23512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.103653] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.103665] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:23520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.103672] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:30 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.103684] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:23528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.103691] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.103706] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:23536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.103713] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:36 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.103725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:23544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.103732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:3 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.103744] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:23552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.103751] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:29 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.103763] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:23560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.103770] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.103783] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:23568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.103790] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:78 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.103803] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:23576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.103809] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.103822] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:23584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.103828] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:16 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.103841] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:23592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.103848] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.104033] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:23600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.104043] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.104056] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:23608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.104063] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:4 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.104075] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:23616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.104082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:35 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.104095] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:23624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.104102] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.104114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:23632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.104123] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:33 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.104136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:23640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.104144] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.104156] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:23648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.104163] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:24:11.891 [2024-07-15 18:09:50.104175] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:23656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.891 [2024-07-15 18:09:50.104182] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:23664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104201] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104213] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:23672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104220] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104239] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:23680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104247] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104259] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:23688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:104 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104280] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:23696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104287] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:72 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104299] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:23704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104306] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:102 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104318] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:23712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104325] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:23720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104355] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:23728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104365] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:22 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104378] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:23736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104384] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:60 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104397] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:23744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104415] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:23752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:23760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104441] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:23768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104460] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104472] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:23776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104478] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:77 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:23784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:125 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104509] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:23792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104516] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104528] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:23800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104535] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:42 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104547] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:23808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104553] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:51 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104566] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:23816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104585] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:23824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104593] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:68 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104607] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:23832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104614] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:99 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104626] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:23840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104634] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:27 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104646] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:23848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104653] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:107 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104665] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:23856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104672] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:58 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104684] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:23864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104691] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104703] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:23872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104710] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:24 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104722] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:23880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104728] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:98 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104740] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:23888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104747] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104759] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:23896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104766] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:38 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104778] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:23904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:56 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104796] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:23912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104804] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:93 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104816] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:23920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:13 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104836] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:23928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104843] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:114 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104855] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:23936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104862] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:103 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104873] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:23944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104893] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:23952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104913] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:119 nsid:1 lba:23960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104920] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:119 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.104932] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:23968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.104940] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:96 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:24:11.892 [2024-07-15 18:09:50.105359] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:23976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.892 [2024-07-15 18:09:50.105370] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:105 nsid:1 lba:23984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.893 [2024-07-15 18:09:50.105391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:105 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:23992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.893 [2024-07-15 18:09:50.105410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:85 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105422] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.893 [2024-07-15 18:09:50.105430] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105443] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:24008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.893 [2024-07-15 18:09:50.105451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105463] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:24016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.893 [2024-07-15 18:09:50.105469] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:88 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105481] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:113 nsid:1 lba:24024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.893 [2024-07-15 18:09:50.105491] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:113 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:118 nsid:1 lba:24032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.893 [2024-07-15 18:09:50.105510] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:118 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105522] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:24040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.893 [2024-07-15 18:09:50.105529] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:52 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105542] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:23104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.105548] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:69 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105560] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:23112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.105567] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105580] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:23120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.105586] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105598] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:23128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.105605] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:5 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105617] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:23136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.105625] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:20 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105637] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:23144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.105644] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:84 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105656] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:23152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.105662] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:90 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105675] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:23160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.105682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:97 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105694] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:23168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.105700] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:49 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105712] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:23176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.105723] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:14 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105735] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:23184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.105742] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105754] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:23192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.105762] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105775] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.105781] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:53 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105793] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:23208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.105800] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105813] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:23216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.105820] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105832] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:24048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.893 [2024-07-15 18:09:50.105839] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:82 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105851] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:24056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.893 [2024-07-15 18:09:50.105858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:48 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105870] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:24064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.893 [2024-07-15 18:09:50.105877] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105889] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:24072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.893 [2024-07-15 18:09:50.105895] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:124 cdw0:0 sqhd:0040 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105908] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.893 [2024-07-15 18:09:50.105914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.105927] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:24088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.893 [2024-07-15 18:09:50.105933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:71 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.106180] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:24096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.893 [2024-07-15 18:09:50.106189] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.106204] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:24104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.893 [2024-07-15 18:09:50.106211] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:39 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.106223] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:24112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.893 [2024-07-15 18:09:50.106235] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:116 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.106247] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:24120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.893 [2024-07-15 18:09:50.106254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:75 cdw0:0 sqhd:0046 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.106266] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:23224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.106273] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:73 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.106286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:23232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.106292] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.106304] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:23240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.106311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.106323] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:23248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.106330] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.106342] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:81 nsid:1 lba:23256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.106350] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:81 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.106362] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:23264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.106368] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:121 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.106381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:23272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.106389] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:117 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.106401] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:23280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.893 [2024-07-15 18:09:50.106408] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:24:11.893 [2024-07-15 18:09:50.106420] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:23288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:87 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106442] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:23296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106449] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106461] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:23304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106468] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:11 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106480] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:23312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106487] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:12 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106499] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106506] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:54 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106518] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:23328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106525] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:86 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106537] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:23336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:9 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:23344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106562] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106574] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:23352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106594] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:23360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106612] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:23368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106619] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106632] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:23376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106638] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:17 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106650] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:23384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106657] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106669] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:23392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106677] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106689] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:23400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106697] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:100 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:23408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106717] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:120 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106729] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:23416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106736] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:50 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106748] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:57 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106767] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:23432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:80 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106786] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:23440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106792] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:2 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106812] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:59 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106824] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:23456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.894 [2024-07-15 18:09:50.106831] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106843] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:23464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.894 [2024-07-15 18:09:50.106850] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:16 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106863] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:23472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.894 [2024-07-15 18:09:50.106870] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106882] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:23480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.894 [2024-07-15 18:09:50.106888] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:78 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.106901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:23488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.894 [2024-07-15 18:09:50.106909] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.107198] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:23496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.894 [2024-07-15 18:09:50.107208] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:29 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.107220] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:23504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.894 [2024-07-15 18:09:50.107232] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:3 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.107244] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:23512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.894 [2024-07-15 18:09:50.107251] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:36 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.107263] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:23520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.894 [2024-07-15 18:09:50.107270] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.107284] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:23528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.894 [2024-07-15 18:09:50.107291] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:30 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.107302] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:23536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.894 [2024-07-15 18:09:50.107310] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.107322] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:23544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.894 [2024-07-15 18:09:50.107328] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.107340] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:23552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.894 [2024-07-15 18:09:50.107347] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.107359] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:23560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.894 [2024-07-15 18:09:50.107366] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.107378] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:23568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.894 [2024-07-15 18:09:50.107387] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:66 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.107400] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:23576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.894 [2024-07-15 18:09:50.107407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.107419] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:23584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.894 [2024-07-15 18:09:50.107426] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.107440] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:23592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.894 [2024-07-15 18:09:50.107447] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:96 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.107459] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:119 nsid:1 lba:23600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.894 [2024-07-15 18:09:50.107466] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:119 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.107478] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:23608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.894 [2024-07-15 18:09:50.107485] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:24:11.894 [2024-07-15 18:09:50.107497] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:23616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.894 [2024-07-15 18:09:50.107504] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107517] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:23624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107525] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:103 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107537] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:23632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107544] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:114 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107556] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:23640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:13 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:23648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:93 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107596] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:23656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107602] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:56 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:23664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107622] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:38 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:23672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107641] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107653] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:23680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107660] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:98 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107674] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:23688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:24 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107695] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:23696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107703] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107715] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:23704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107722] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:58 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107734] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:23712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107741] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:107 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107753] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:23720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107760] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:27 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107772] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:23728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:99 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107791] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:23736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107798] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:68 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:23744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107817] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:23752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107836] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:51 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107848] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:23760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107855] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:42 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:23768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107874] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107886] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:23776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107893] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:125 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107906] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:23784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107914] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:77 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107927] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:23792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107935] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107949] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:23800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107956] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:23808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.107988] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:23816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.107995] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.108008] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:23824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.108015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:60 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.108027] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:23832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.108034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:22 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.108046] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:23840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.108053] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.108065] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:23848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.108072] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.108085] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:23856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.108091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:102 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.108103] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:23864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.108110] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:72 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.108122] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:23872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.108129] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:104 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.108141] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:23880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.108149] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.108161] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:23888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.108168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.108180] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:23896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.108187] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.108656] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:23904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.108669] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.108683] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:23912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.108690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.108702] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:23920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.108709] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.108721] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:23928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.108727] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:33 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.108740] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:23936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.108746] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:24:11.895 [2024-07-15 18:09:50.108758] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:23944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.895 [2024-07-15 18:09:50.108765] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:35 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.108777] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:4 nsid:1 lba:23952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.896 [2024-07-15 18:09:50.108784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:4 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.108796] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:23960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.896 [2024-07-15 18:09:50.108803] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.108815] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:71 nsid:1 lba:23968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.896 [2024-07-15 18:09:50.108822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:71 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.108834] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:23976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.896 [2024-07-15 18:09:50.108841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.108856] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:124 nsid:1 lba:23984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.896 [2024-07-15 18:09:50.108863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:124 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.108875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:89 nsid:1 lba:23992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.896 [2024-07-15 18:09:50.108882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.108894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:24000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.896 [2024-07-15 18:09:50.108901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:48 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.108914] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:24008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.896 [2024-07-15 18:09:50.108921] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:82 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.108933] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:24016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.896 [2024-07-15 18:09:50.108940] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.108952] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:24024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.896 [2024-07-15 18:09:50.108959] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.108971] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:24032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.896 [2024-07-15 18:09:50.108978] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:53 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.108990] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:24040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.896 [2024-07-15 18:09:50.108997] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:23104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.896 [2024-07-15 18:09:50.109016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109028] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:23112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.896 [2024-07-15 18:09:50.109035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:14 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:23120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.896 [2024-07-15 18:09:50.109054] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:49 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109066] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:23128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.896 [2024-07-15 18:09:50.109073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:97 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:23136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.896 [2024-07-15 18:09:50.109094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:90 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109106] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:23144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.896 [2024-07-15 18:09:50.109113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:84 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109126] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:23152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.896 [2024-07-15 18:09:50.109134] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:20 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109146] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:23160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.896 [2024-07-15 18:09:50.109152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:5 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109165] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:23168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.896 [2024-07-15 18:09:50.109171] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109183] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:23176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.896 [2024-07-15 18:09:50.109190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109202] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:23184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.896 [2024-07-15 18:09:50.109209] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:69 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109221] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:23192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.896 [2024-07-15 18:09:50.109233] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:52 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109246] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:23200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.896 [2024-07-15 18:09:50.109254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:118 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109266] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:23208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.896 [2024-07-15 18:09:50.109273] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:113 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:23216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.896 [2024-07-15 18:09:50.109293] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:88 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109306] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:24048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.896 [2024-07-15 18:09:50.109313] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109325] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.896 [2024-07-15 18:09:50.109334] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:24064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.896 [2024-07-15 18:09:50.109354] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:85 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109648] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:105 nsid:1 lba:24072 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.896 [2024-07-15 18:09:50.109658] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:105 cdw0:0 sqhd:0040 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109672] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:108 nsid:1 lba:24080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.896 [2024-07-15 18:09:50.109680] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:108 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109692] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:24088 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.896 [2024-07-15 18:09:50.109699] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0042 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109712] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:24096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.896 [2024-07-15 18:09:50.109718] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:78 cdw0:0 sqhd:0043 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109731] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:24104 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.896 [2024-07-15 18:09:50.109738] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:31 cdw0:0 sqhd:0044 p:0 m:0 dnr:0 00:24:11.896 [2024-07-15 18:09:50.109750] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:24112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.897 [2024-07-15 18:09:50.109757] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:16 cdw0:0 sqhd:0045 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.109769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:24120 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.897 [2024-07-15 18:09:50.109776] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:19 cdw0:0 sqhd:0046 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.109788] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:23224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.109795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:59 cdw0:0 sqhd:0047 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.109807] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:23232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.109815] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:2 cdw0:0 sqhd:0048 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.109827] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:23240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.109834] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:80 cdw0:0 sqhd:0049 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.109847] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:23248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.109855] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:57 cdw0:0 sqhd:004a p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.109868] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:23256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.109875] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:50 cdw0:0 sqhd:004b p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.109887] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:23264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.109894] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:120 cdw0:0 sqhd:004c p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.109906] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:23272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.109913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:100 cdw0:0 sqhd:004d p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.109926] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:23280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.109933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:004e p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.109945] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:23288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.109951] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:44 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.109964] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:23296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.109971] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:17 cdw0:0 sqhd:0050 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.109983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:23304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.109990] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:23 cdw0:0 sqhd:0051 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.110006] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:23312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.110012] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:62 cdw0:0 sqhd:0052 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.110025] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:23320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.110032] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:95 cdw0:0 sqhd:0053 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.110045] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:23328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.110053] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:0054 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.110067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:23336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.110074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:9 cdw0:0 sqhd:0055 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.110086] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:23344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.110094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:86 cdw0:0 sqhd:0056 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.110108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:23352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.110115] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:54 cdw0:0 sqhd:0057 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.110128] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:23360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.110135] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:12 cdw0:0 sqhd:0058 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.110148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:23368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.110155] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:11 cdw0:0 sqhd:0059 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.110168] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:23376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.110177] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:112 cdw0:0 sqhd:005a p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.110190] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:23384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.110198] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:87 cdw0:0 sqhd:005b p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.110211] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:23392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.110218] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:005c p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.110236] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:23400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.110243] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:117 cdw0:0 sqhd:005d p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.110256] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:23408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.110263] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:121 cdw0:0 sqhd:005e p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.114136] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:81 nsid:1 lba:23416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.114145] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:81 cdw0:0 sqhd:005f p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.114159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:23424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.114167] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:7 cdw0:0 sqhd:0060 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.114180] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:23432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.114188] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.114202] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:23440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.114209] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:0062 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.114223] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:23448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.114242] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:73 cdw0:0 sqhd:0063 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.114255] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:23456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.897 [2024-07-15 18:09:50.114261] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:75 cdw0:0 sqhd:0064 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.114274] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:23464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.897 [2024-07-15 18:09:50.114280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:116 cdw0:0 sqhd:0065 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.114425] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:23472 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.897 [2024-07-15 18:09:50.114435] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:39 cdw0:0 sqhd:0066 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.114461] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:23480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.897 [2024-07-15 18:09:50.114469] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0067 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.114484] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:23488 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.897 [2024-07-15 18:09:50.114491] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0068 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.114506] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:23496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.897 [2024-07-15 18:09:50.114513] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0069 p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.114527] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:23504 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.897 [2024-07-15 18:09:50.114534] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:006a p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.114549] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:23512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.897 [2024-07-15 18:09:50.114556] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:104 cdw0:0 sqhd:006b p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.114570] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:23520 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.897 [2024-07-15 18:09:50.114577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:72 cdw0:0 sqhd:006c p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.114592] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:23528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.897 [2024-07-15 18:09:50.114599] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:102 cdw0:0 sqhd:006d p:0 m:0 dnr:0 00:24:11.897 [2024-07-15 18:09:50.114613] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:32 nsid:1 lba:23536 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.897 [2024-07-15 18:09:50.114620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:32 cdw0:0 sqhd:006e p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.114639] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:23544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.114646] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:006f p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.114661] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:23552 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.114667] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:22 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.114682] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:23560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.114689] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:60 cdw0:0 sqhd:0071 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.114703] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:111 nsid:1 lba:23568 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.114710] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:111 cdw0:0 sqhd:0072 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.114725] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:23576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.114732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:0073 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.114746] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:23584 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.114753] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:0074 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.114767] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:23592 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.114774] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:0075 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.114789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:23600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.114796] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:77 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.114810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:125 nsid:1 lba:23608 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.114817] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:125 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.114831] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:23616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.114838] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.114852] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:23624 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.114859] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:42 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.114874] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:23632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.114880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:51 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.114895] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:23640 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.114903] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:28 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.114918] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:23648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.114924] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:68 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.114939] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:99 nsid:1 lba:23656 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.114946] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:99 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.114960] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:23664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.114967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:27 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.114981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:107 nsid:1 lba:23672 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.114988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:107 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115003] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:23680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:58 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115024] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:23688 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115031] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:18 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115046] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:23696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115053] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:24 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:98 nsid:1 lba:23704 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:98 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:23712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115096] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115110] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:23720 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115117] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:38 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:23728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:56 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115153] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:23736 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115161] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:93 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115175] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:23744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115182] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:13 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115196] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:114 nsid:1 lba:23752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115203] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:114 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115218] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:23760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115231] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:103 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115247] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:74 nsid:1 lba:23768 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:74 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115269] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:23776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115276] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115290] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:119 nsid:1 lba:23784 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:119 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115313] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:23792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115320] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:96 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115334] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:23800 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115341] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:23808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115362] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:66 nsid:1 lba:23816 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115384] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:66 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115398] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:92 nsid:1 lba:23824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115405] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:92 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115420] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:23832 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115442] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:23840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115449] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115464] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:23848 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115471] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115486] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:23856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115493] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:30 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115507] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:23864 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115514] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:83 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:24:11.898 [2024-07-15 18:09:50.115528] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:23872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.898 [2024-07-15 18:09:50.115535] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:36 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:23880 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115556] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:3 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115571] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:23888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115578] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:29 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115592] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:23896 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115599] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:85 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115613] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115635] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:23912 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115642] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115657] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:88 nsid:1 lba:23920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115663] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:88 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115678] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:113 nsid:1 lba:23928 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115685] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:113 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115701] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:118 nsid:1 lba:23936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115708] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:118 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115722] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:52 nsid:1 lba:23944 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115729] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:52 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115745] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:23952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115751] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:69 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:23960 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115787] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:94 nsid:1 lba:23968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115794] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:94 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115809] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:5 nsid:1 lba:23976 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115816] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:5 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115830] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:23984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115837] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:20 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115852] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:23992 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115859] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:84 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115873] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:90 nsid:1 lba:24000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:90 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:24008 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:97 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115916] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:24016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115922] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:49 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115937] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:24024 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115944] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:14 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:24032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.115982] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:24040 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.115988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:26 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.116003] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:23104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.899 [2024-07-15 18:09:50.116010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:53 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.116024] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:23112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.899 [2024-07-15 18:09:50.116031] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.116045] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:23120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.899 [2024-07-15 18:09:50.116052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:8 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.116067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:82 nsid:1 lba:23128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.899 [2024-07-15 18:09:50.116074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:82 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.116089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:23136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.899 [2024-07-15 18:09:50.116096] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:48 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.116110] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:23144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.899 [2024-07-15 18:09:50.116117] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:89 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.116132] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:124 nsid:1 lba:23152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.899 [2024-07-15 18:09:50.116138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:124 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.116153] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:23160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.899 [2024-07-15 18:09:50.116159] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:0 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.116174] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:23168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.899 [2024-07-15 18:09:50.116181] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:71 cdw0:0 sqhd:0036 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.116195] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:23176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.899 [2024-07-15 18:09:50.116202] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:0037 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.116217] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:23184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.899 [2024-07-15 18:09:50.116228] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:4 cdw0:0 sqhd:0038 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.116244] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:23192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.899 [2024-07-15 18:09:50.116250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:35 cdw0:0 sqhd:0039 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.116265] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:23200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.899 [2024-07-15 18:09:50.116272] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:34 cdw0:0 sqhd:003a p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.116286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:23208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.899 [2024-07-15 18:09:50.116293] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:33 cdw0:0 sqhd:003b p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.116308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:23216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.899 [2024-07-15 18:09:50.116315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:003c p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.116329] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:24048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.116336] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:003d p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.116351] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:24056 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.116358] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:46 cdw0:0 sqhd:003e p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:09:50.116521] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:116 nsid:1 lba:24064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.899 [2024-07-15 18:09:50.116529] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:116 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:10:02.903321] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:66432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.899 [2024-07-15 18:10:02.903362] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:88 cdw0:0 sqhd:0076 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:10:02.903395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:66424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.899 [2024-07-15 18:10:02.903404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:87 cdw0:0 sqhd:0077 p:0 m:0 dnr:0 00:24:11.899 [2024-07-15 18:10:02.903418] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:126 nsid:1 lba:66464 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.903425] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:126 cdw0:0 sqhd:0078 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.903438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:47 nsid:1 lba:66480 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.903445] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:47 cdw0:0 sqhd:0079 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.903457] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:68 nsid:1 lba:66496 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.903464] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:68 cdw0:0 sqhd:007a p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.903481] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:66512 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.903488] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:81 cdw0:0 sqhd:007b p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.903500] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:66528 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.903507] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:73 cdw0:0 sqhd:007c p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.903519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:91 nsid:1 lba:66544 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.903526] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:91 cdw0:0 sqhd:007d p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.903538] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:66560 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.903545] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:115 cdw0:0 sqhd:007e p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.903557] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:66576 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.903564] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:1 cdw0:0 sqhd:007f p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.903578] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:66448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:24:11.900 [2024-07-15 18:10:02.903584] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:37 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904297] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:66600 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:78 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904326] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:66616 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904334] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:42 cdw0:0 sqhd:0002 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904346] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:66632 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904353] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:25 cdw0:0 sqhd:0003 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:60 nsid:1 lba:66648 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904373] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:60 cdw0:0 sqhd:0004 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904385] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:106 nsid:1 lba:66664 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904392] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:106 cdw0:0 sqhd:0005 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:66680 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:40 cdw0:0 sqhd:0006 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904428] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:109 nsid:1 lba:66696 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904435] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:109 cdw0:0 sqhd:0007 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904447] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:120 nsid:1 lba:66712 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904454] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:120 cdw0:0 sqhd:0008 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904466] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:58 nsid:1 lba:66728 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904473] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:58 cdw0:0 sqhd:0009 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904486] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:103 nsid:1 lba:66744 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904493] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:103 cdw0:0 sqhd:000a p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904505] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:64 nsid:1 lba:66760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:64 cdw0:0 sqhd:000b p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904524] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:66776 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904530] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:35 cdw0:0 sqhd:000c p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904543] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:97 nsid:1 lba:66792 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904550] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:97 cdw0:0 sqhd:000d p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904562] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:61 nsid:1 lba:66808 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904568] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:61 cdw0:0 sqhd:000e p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904581] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:66824 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:24 cdw0:0 sqhd:000f p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:66840 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904607] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:39 cdw0:0 sqhd:0010 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904619] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:66856 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904626] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:10 cdw0:0 sqhd:0011 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904638] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:122 nsid:1 lba:66872 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904645] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:122 cdw0:0 sqhd:0012 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904657] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:96 nsid:1 lba:66888 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904665] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:96 cdw0:0 sqhd:0013 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904678] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:66904 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904684] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:2 cdw0:0 sqhd:0014 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904697] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:119 nsid:1 lba:66920 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:119 cdw0:0 sqhd:0015 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.904716] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:51 nsid:1 lba:66936 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.904723] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:51 cdw0:0 sqhd:0016 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.905735] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:66952 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.905748] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:12 cdw0:0 sqhd:0017 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.905762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:104 nsid:1 lba:66968 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.905769] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:104 cdw0:0 sqhd:0018 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.905781] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:66984 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.905788] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:15 cdw0:0 sqhd:0019 p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.905801] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:67000 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.905808] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:79 cdw0:0 sqhd:001a p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.905820] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:67016 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.905827] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:110 cdw0:0 sqhd:001b p:0 m:0 dnr:0 00:24:11.900 [2024-07-15 18:10:02.905840] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:67032 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.900 [2024-07-15 18:10:02.905846] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:3 cdw0:0 sqhd:001c p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.905859] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:67048 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.905866] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:82 cdw0:0 sqhd:001d p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.905880] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:6 nsid:1 lba:67064 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.905887] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:6 cdw0:0 sqhd:001e p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.905899] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:67 nsid:1 lba:67080 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.905908] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:67 cdw0:0 sqhd:001f p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.905921] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:57 nsid:1 lba:67096 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.905928] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:57 cdw0:0 sqhd:0020 p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.905941] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:67112 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.905947] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:75 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.905960] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:65 nsid:1 lba:67128 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.905967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:65 cdw0:0 sqhd:0022 p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.905979] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:67144 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.905985] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:45 cdw0:0 sqhd:0023 p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.905998] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:123 nsid:1 lba:67160 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.906005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:123 cdw0:0 sqhd:0024 p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.906017] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:55 nsid:1 lba:67176 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.906024] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:55 cdw0:0 sqhd:0025 p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.906036] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:67192 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.906042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:21 cdw0:0 sqhd:0026 p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.906055] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:67208 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.906061] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:41 cdw0:0 sqhd:0027 p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.906238] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:67224 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.906248] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:76 cdw0:0 sqhd:0028 p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.906261] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:67240 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.906269] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:17 cdw0:0 sqhd:0029 p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.906281] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:49 nsid:1 lba:67256 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.906288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:49 cdw0:0 sqhd:002a p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.906300] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:67272 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.906307] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:43 cdw0:0 sqhd:002b p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.906321] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:63 nsid:1 lba:67288 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.906328] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:63 cdw0:0 sqhd:002c p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.906340] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:53 nsid:1 lba:67304 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.906346] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:53 cdw0:0 sqhd:002d p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.906359] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:102 nsid:1 lba:67320 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.906365] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:102 cdw0:0 sqhd:002e p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.906378] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:48 nsid:1 lba:67336 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.906385] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:48 cdw0:0 sqhd:002f p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.906397] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:67352 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.906403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:101 cdw0:0 sqhd:0030 p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.906415] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:56 nsid:1 lba:67368 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.906422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:56 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.906435] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:67384 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.906441] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:29 cdw0:0 sqhd:0032 p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.906453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:93 nsid:1 lba:67400 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.906460] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:93 cdw0:0 sqhd:0033 p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.906472] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:70 nsid:1 lba:67416 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.906478] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:70 cdw0:0 sqhd:0034 p:0 m:0 dnr:0 00:24:11.901 [2024-07-15 18:10:02.906491] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:67432 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:24:11.901 [2024-07-15 18:10:02.906497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ASYMMETRIC ACCESS INACCESSIBLE (03/02) qid:1 cid:36 cdw0:0 sqhd:0035 p:0 m:0 dnr:0 00:24:11.901 Received shutdown signal, test time was about 27.029419 seconds 00:24:11.901 00:24:11.901 Latency(us) 00:24:11.901 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:24:11.901 Job: Nvme0n1 (Core Mask 0x4, workload: verify, depth: 128, IO size: 4096) 00:24:11.901 Verification LBA range: start 0x0 length 0x4000 00:24:11.901 Nvme0n1 : 27.03 10300.73 40.24 0.00 0.00 12384.56 407.82 3078254.41 00:24:11.901 =================================================================================================================== 00:24:11.901 Total : 10300.73 40.24 0.00 0.00 12384.56 407.82 3078254.41 00:24:11.901 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@143 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:24:11.901 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@145 -- # trap - SIGINT SIGTERM EXIT 00:24:11.901 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@147 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/try.txt 00:24:11.901 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- host/multipath_status.sh@148 -- # nvmftestfini 00:24:11.901 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@488 -- # nvmfcleanup 00:24:11.901 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@117 -- # sync 00:24:11.902 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:24:11.902 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@120 -- # set +e 00:24:11.902 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@121 -- # for i in {1..20} 00:24:11.902 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:24:11.902 rmmod nvme_tcp 00:24:11.902 rmmod nvme_fabrics 00:24:11.902 rmmod nvme_keyring 00:24:11.902 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:24:11.902 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@124 -- # set -e 00:24:11.902 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@125 -- # return 0 00:24:11.902 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@489 -- # '[' -n 696647 ']' 00:24:11.902 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@490 -- # killprocess 696647 00:24:11.902 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@948 -- # '[' -z 696647 ']' 00:24:11.902 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@952 -- # kill -0 696647 00:24:11.902 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@953 -- # uname 00:24:11.902 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:24:11.902 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 696647 00:24:12.160 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:24:12.160 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:24:12.160 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@966 -- # echo 'killing process with pid 696647' 00:24:12.160 killing process with pid 696647 00:24:12.160 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@967 -- # kill 696647 00:24:12.160 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@972 -- # wait 696647 00:24:12.160 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:24:12.160 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:24:12.160 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:24:12.160 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:24:12.160 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@278 -- # remove_spdk_ns 00:24:12.160 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:12.160 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:24:12.160 18:10:05 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:14.693 18:10:07 nvmf_tcp.nvmf_host_multipath_status -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:24:14.693 00:24:14.693 real 0m39.273s 00:24:14.693 user 1m46.021s 00:24:14.693 sys 0m10.512s 00:24:14.693 18:10:07 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@1124 -- # xtrace_disable 00:24:14.693 18:10:07 nvmf_tcp.nvmf_host_multipath_status -- common/autotest_common.sh@10 -- # set +x 00:24:14.693 ************************************ 00:24:14.693 END TEST nvmf_host_multipath_status 00:24:14.693 ************************************ 00:24:14.693 18:10:07 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:24:14.693 18:10:07 nvmf_tcp -- nvmf/nvmf.sh@103 -- # run_test nvmf_discovery_remove_ifc /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery_remove_ifc.sh --transport=tcp 00:24:14.693 18:10:07 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:24:14.694 18:10:07 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:24:14.694 18:10:07 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:24:14.694 ************************************ 00:24:14.694 START TEST nvmf_discovery_remove_ifc 00:24:14.694 ************************************ 00:24:14.694 18:10:07 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/discovery_remove_ifc.sh --transport=tcp 00:24:14.694 * Looking for test storage... 00:24:14.694 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:24:14.694 18:10:07 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:24:14.694 18:10:07 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@7 -- # uname -s 00:24:14.694 18:10:07 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:24:14.694 18:10:07 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:24:14.694 18:10:07 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:24:14.694 18:10:07 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:24:14.694 18:10:07 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:24:14.694 18:10:07 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:24:14.694 18:10:07 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:24:14.694 18:10:07 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:24:14.694 18:10:07 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:24:14.694 18:10:07 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:24:14.694 18:10:07 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:24:14.694 18:10:07 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:24:14.694 18:10:07 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:24:14.694 18:10:07 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:24:14.694 18:10:07 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:24:14.694 18:10:07 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- paths/export.sh@5 -- # export PATH 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@47 -- # : 0 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@51 -- # have_pci_nics=0 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@14 -- # '[' tcp == rdma ']' 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@19 -- # discovery_port=8009 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@20 -- # discovery_nqn=nqn.2014-08.org.nvmexpress.discovery 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@23 -- # nqn=nqn.2016-06.io.spdk:cnode 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@25 -- # host_nqn=nqn.2021-12.io.spdk:test 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@26 -- # host_sock=/tmp/host.sock 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@39 -- # nvmftestinit 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@448 -- # prepare_net_devs 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@410 -- # local -g is_hw=no 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@412 -- # remove_spdk_ns 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@285 -- # xtrace_disable 00:24:14.694 18:10:08 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@291 -- # pci_devs=() 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@291 -- # local -a pci_devs 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@292 -- # pci_net_devs=() 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@293 -- # pci_drivers=() 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@293 -- # local -A pci_drivers 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@295 -- # net_devs=() 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@295 -- # local -ga net_devs 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@296 -- # e810=() 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@296 -- # local -ga e810 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@297 -- # x722=() 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@297 -- # local -ga x722 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@298 -- # mlx=() 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@298 -- # local -ga mlx 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:24:19.998 Found 0000:86:00.0 (0x8086 - 0x159b) 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:24:19.998 Found 0000:86:00.1 (0x8086 - 0x159b) 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@390 -- # [[ up == up ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:24:19.998 Found net devices under 0000:86:00.0: cvl_0_0 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@390 -- # [[ up == up ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:24:19.998 Found net devices under 0000:86:00.1: cvl_0_1 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@414 -- # is_hw=yes 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:24:19.998 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:24:19.999 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:24:19.999 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:24:19.999 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:24:19.999 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:24:19.999 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:24:19.999 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:24:19.999 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:24:19.999 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:24:19.999 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:24:19.999 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:24:19.999 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:24:19.999 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:24:19.999 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:24:19.999 18:10:12 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:24:19.999 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:24:19.999 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.192 ms 00:24:19.999 00:24:19.999 --- 10.0.0.2 ping statistics --- 00:24:19.999 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:19.999 rtt min/avg/max/mdev = 0.192/0.192/0.192/0.000 ms 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:24:19.999 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:24:19.999 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.167 ms 00:24:19.999 00:24:19.999 --- 10.0.0.1 ping statistics --- 00:24:19.999 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:19.999 rtt min/avg/max/mdev = 0.167/0.167/0.167/0.000 ms 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@422 -- # return 0 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@40 -- # nvmfappstart -m 0x2 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@722 -- # xtrace_disable 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@481 -- # nvmfpid=705295 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@482 -- # waitforlisten 705295 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@829 -- # '[' -z 705295 ']' 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@834 -- # local max_retries=100 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:19.999 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0x2 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@838 -- # xtrace_disable 00:24:19.999 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:19.999 [2024-07-15 18:10:13.153053] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:24:19.999 [2024-07-15 18:10:13.153103] [ DPDK EAL parameters: nvmf -c 0x2 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:24:19.999 EAL: No free 2048 kB hugepages reported on node 1 00:24:19.999 [2024-07-15 18:10:13.211940] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:19.999 [2024-07-15 18:10:13.293737] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:24:19.999 [2024-07-15 18:10:13.293771] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:24:19.999 [2024-07-15 18:10:13.293778] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:24:19.999 [2024-07-15 18:10:13.293784] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:24:19.999 [2024-07-15 18:10:13.293788] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:24:19.999 [2024-07-15 18:10:13.293828] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:24:20.257 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:24:20.257 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@862 -- # return 0 00:24:20.257 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:24:20.257 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@728 -- # xtrace_disable 00:24:20.257 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:20.257 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:24:20.257 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@43 -- # rpc_cmd 00:24:20.257 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:20.257 18:10:13 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:20.516 [2024-07-15 18:10:13.985250] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:24:20.516 [2024-07-15 18:10:13.993359] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 8009 *** 00:24:20.516 null0 00:24:20.516 [2024-07-15 18:10:14.025378] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:24:20.516 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:20.516 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@59 -- # hostpid=705449 00:24:20.516 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@58 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -m 0x1 -r /tmp/host.sock --wait-for-rpc -L bdev_nvme 00:24:20.516 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@60 -- # waitforlisten 705449 /tmp/host.sock 00:24:20.516 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@829 -- # '[' -z 705449 ']' 00:24:20.516 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@833 -- # local rpc_addr=/tmp/host.sock 00:24:20.516 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@834 -- # local max_retries=100 00:24:20.516 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock...' 00:24:20.516 Waiting for process to start up and listen on UNIX domain socket /tmp/host.sock... 00:24:20.516 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@838 -- # xtrace_disable 00:24:20.516 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:20.516 [2024-07-15 18:10:14.078726] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:24:20.516 [2024-07-15 18:10:14.078765] [ DPDK EAL parameters: nvmf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid705449 ] 00:24:20.516 EAL: No free 2048 kB hugepages reported on node 1 00:24:20.516 [2024-07-15 18:10:14.132923] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:24:20.516 [2024-07-15 18:10:14.213245] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:24:21.452 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:24:21.452 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@862 -- # return 0 00:24:21.452 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@62 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; killprocess $hostpid; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:24:21.452 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@65 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_set_options -e 1 00:24:21.452 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:21.452 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:21.452 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:21.452 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@66 -- # rpc_cmd -s /tmp/host.sock framework_start_init 00:24:21.452 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:21.452 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:21.452 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:21.452 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@69 -- # rpc_cmd -s /tmp/host.sock bdev_nvme_start_discovery -b nvme -t tcp -a 10.0.0.2 -s 8009 -f ipv4 -q nqn.2021-12.io.spdk:test --ctrlr-loss-timeout-sec 2 --reconnect-delay-sec 1 --fast-io-fail-timeout-sec 1 --wait-for-attach 00:24:21.452 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:21.452 18:10:14 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:22.387 [2024-07-15 18:10:16.033719] bdev_nvme.c:6983:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:24:22.387 [2024-07-15 18:10:16.033743] bdev_nvme.c:7063:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:24:22.387 [2024-07-15 18:10:16.033755] bdev_nvme.c:6946:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:24:22.645 [2024-07-15 18:10:16.160149] bdev_nvme.c:6912:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 new subsystem nvme0 00:24:22.645 [2024-07-15 18:10:16.216976] bdev_nvme.c:7773:bdev_nvme_readv: *DEBUG*: read 8 blocks with offset 0 00:24:22.645 [2024-07-15 18:10:16.217019] bdev_nvme.c:7773:bdev_nvme_readv: *DEBUG*: read 1 blocks with offset 0 00:24:22.645 [2024-07-15 18:10:16.217039] bdev_nvme.c:7773:bdev_nvme_readv: *DEBUG*: read 64 blocks with offset 0 00:24:22.645 [2024-07-15 18:10:16.217051] bdev_nvme.c:6802:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme0 done 00:24:22.645 [2024-07-15 18:10:16.217069] bdev_nvme.c:6761:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:24:22.645 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:22.645 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@72 -- # wait_for_bdev nvme0n1 00:24:22.645 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:24:22.645 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:24:22.645 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:24:22.645 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:24:22.645 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:22.645 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:24:22.645 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:22.645 [2024-07-15 18:10:16.222409] bdev_nvme.c:1617:bdev_nvme_disconnected_qpair_cb: *DEBUG*: qpair 0x19cde30 was disconnected and freed. delete nvme_qpair. 00:24:22.645 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:22.645 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != \n\v\m\e\0\n\1 ]] 00:24:22.646 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@75 -- # ip netns exec cvl_0_0_ns_spdk ip addr del 10.0.0.2/24 dev cvl_0_0 00:24:22.646 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@76 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 down 00:24:22.646 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@79 -- # wait_for_bdev '' 00:24:22.646 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:24:22.646 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:24:22.646 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:24:22.646 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:24:22.646 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:22.646 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:24:22.646 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:22.905 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:22.905 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:24:22.905 18:10:16 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:24:23.840 18:10:17 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:24:23.840 18:10:17 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:24:23.840 18:10:17 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:24:23.840 18:10:17 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:24:23.840 18:10:17 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:23.840 18:10:17 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:23.840 18:10:17 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:24:23.840 18:10:17 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:23.840 18:10:17 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:24:23.840 18:10:17 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:24:24.775 18:10:18 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:24:24.775 18:10:18 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:24:24.775 18:10:18 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:24:24.775 18:10:18 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:24:24.775 18:10:18 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:24.775 18:10:18 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:24:24.775 18:10:18 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:24.775 18:10:18 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:25.034 18:10:18 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:24:25.034 18:10:18 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:24:25.970 18:10:19 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:24:25.970 18:10:19 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:24:25.970 18:10:19 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:24:25.970 18:10:19 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:25.970 18:10:19 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:24:25.970 18:10:19 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:25.970 18:10:19 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:24:25.970 18:10:19 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:25.970 18:10:19 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:24:25.970 18:10:19 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:24:26.906 18:10:20 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:24:26.906 18:10:20 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:24:26.906 18:10:20 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:24:26.906 18:10:20 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:24:26.906 18:10:20 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:26.906 18:10:20 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:24:26.906 18:10:20 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:26.906 18:10:20 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:26.906 18:10:20 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:24:26.906 18:10:20 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:24:28.279 18:10:21 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:24:28.279 18:10:21 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:24:28.279 18:10:21 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:24:28.279 18:10:21 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:24:28.279 18:10:21 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:28.279 18:10:21 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:24:28.279 18:10:21 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:28.279 18:10:21 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:28.279 [2024-07-15 18:10:21.658427] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 428:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 110: Connection timed out 00:24:28.279 [2024-07-15 18:10:21.658462] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:24:28.279 [2024-07-15 18:10:21.658473] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:28.279 [2024-07-15 18:10:21.658498] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:24:28.279 [2024-07-15 18:10:21.658505] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:28.279 [2024-07-15 18:10:21.658512] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:24:28.279 [2024-07-15 18:10:21.658518] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:28.279 [2024-07-15 18:10:21.658526] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:24:28.279 [2024-07-15 18:10:21.658532] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:28.279 [2024-07-15 18:10:21.658540] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:4 nsid:0 cdw10:00000000 cdw11:00000000 00:24:28.279 [2024-07-15 18:10:21.658546] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:28.279 [2024-07-15 18:10:21.658552] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x19946c0 is same with the state(5) to be set 00:24:28.279 [2024-07-15 18:10:21.668448] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x19946c0 (9): Bad file descriptor 00:24:28.279 18:10:21 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:24:28.279 [2024-07-15 18:10:21.678486] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:24:28.279 18:10:21 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:24:29.213 18:10:22 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:24:29.213 18:10:22 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:24:29.213 18:10:22 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:24:29.213 18:10:22 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:24:29.213 18:10:22 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:29.213 18:10:22 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:29.213 18:10:22 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:24:29.213 [2024-07-15 18:10:22.690244] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 110 00:24:29.213 [2024-07-15 18:10:22.690282] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x19946c0 with addr=10.0.0.2, port=4420 00:24:29.213 [2024-07-15 18:10:22.690295] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x19946c0 is same with the state(5) to be set 00:24:29.213 [2024-07-15 18:10:22.690322] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x19946c0 (9): Bad file descriptor 00:24:29.213 [2024-07-15 18:10:22.690729] bdev_nvme.c:2899:bdev_nvme_failover_ctrlr_unsafe: *NOTICE*: Unable to perform failover, already in progress. 00:24:29.213 [2024-07-15 18:10:22.690754] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:24:29.213 [2024-07-15 18:10:22.690763] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:24:29.213 [2024-07-15 18:10:22.690774] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:24:29.213 [2024-07-15 18:10:22.690793] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:24:29.213 [2024-07-15 18:10:22.690803] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] resetting controller 00:24:29.213 18:10:22 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:29.213 18:10:22 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme0n1 != '' ]] 00:24:29.213 18:10:22 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:24:30.148 [2024-07-15 18:10:23.693282] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:24:30.148 [2024-07-15 18:10:23.693304] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:24:30.148 [2024-07-15 18:10:23.693310] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode0] controller reinitialization failed 00:24:30.148 [2024-07-15 18:10:23.693316] nvme_ctrlr.c:1094:nvme_ctrlr_fail: *NOTICE*: [nqn.2016-06.io.spdk:cnode0] already in failed state 00:24:30.148 [2024-07-15 18:10:23.693328] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:24:30.148 [2024-07-15 18:10:23.693345] bdev_nvme.c:6734:remove_discovery_entry: *INFO*: Discovery[10.0.0.2:8009] Remove discovery entry: nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 00:24:30.148 [2024-07-15 18:10:23.693364] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:0 nsid:0 cdw10:00000000 cdw11:00000000 00:24:30.148 [2024-07-15 18:10:23.693372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:30.148 [2024-07-15 18:10:23.693381] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:1 nsid:0 cdw10:00000000 cdw11:00000000 00:24:30.148 [2024-07-15 18:10:23.693387] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:30.148 [2024-07-15 18:10:23.693394] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:2 nsid:0 cdw10:00000000 cdw11:00000000 00:24:30.148 [2024-07-15 18:10:23.693401] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:30.148 [2024-07-15 18:10:23.693408] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: ASYNC EVENT REQUEST (0c) qid:0 cid:3 nsid:0 cdw10:00000000 cdw11:00000000 00:24:30.148 [2024-07-15 18:10:23.693414] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:30.148 [2024-07-15 18:10:23.693421] nvme_qpair.c: 223:nvme_admin_qpair_print_command: *NOTICE*: KEEP ALIVE (18) qid:0 cid:4 nsid:0 cdw10:00000000 cdw11:00000000 00:24:30.148 [2024-07-15 18:10:23.693428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:0 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:24:30.148 [2024-07-15 18:10:23.693434] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2014-08.org.nvmexpress.discovery] in failed state. 00:24:30.148 [2024-07-15 18:10:23.693579] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1993b40 (9): Bad file descriptor 00:24:30.148 [2024-07-15 18:10:23.694590] nvme_fabric.c: 214:nvme_fabric_prop_get_cmd_async: *ERROR*: Failed to send Property Get fabrics command 00:24:30.148 [2024-07-15 18:10:23.694599] nvme_ctrlr.c:1213:nvme_ctrlr_shutdown_async: *ERROR*: [nqn.2014-08.org.nvmexpress.discovery] Failed to read the CC register 00:24:30.148 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:24:30.148 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:24:30.148 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:24:30.148 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:30.148 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:24:30.148 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:30.148 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:24:30.148 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:30.148 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ '' != '' ]] 00:24:30.148 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@82 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:24:30.148 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@83 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:24:30.148 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@86 -- # wait_for_bdev nvme1n1 00:24:30.148 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:24:30.148 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:24:30.148 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:24:30.148 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:24:30.148 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:30.148 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:24:30.148 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:30.148 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:30.407 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ '' != \n\v\m\e\1\n\1 ]] 00:24:30.407 18:10:23 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:24:31.342 18:10:24 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:24:31.342 18:10:24 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:24:31.342 18:10:24 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:24:31.342 18:10:24 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:31.342 18:10:24 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:24:31.342 18:10:24 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:31.342 18:10:24 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:24:31.343 18:10:24 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:31.343 18:10:24 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ '' != \n\v\m\e\1\n\1 ]] 00:24:31.343 18:10:24 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@34 -- # sleep 1 00:24:32.279 [2024-07-15 18:10:25.748312] bdev_nvme.c:6983:discovery_attach_cb: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr attached 00:24:32.279 [2024-07-15 18:10:25.748329] bdev_nvme.c:7063:discovery_poller: *INFO*: Discovery[10.0.0.2:8009] discovery ctrlr connected 00:24:32.279 [2024-07-15 18:10:25.748344] bdev_nvme.c:6946:get_discovery_log_page: *INFO*: Discovery[10.0.0.2:8009] sent discovery log page command 00:24:32.279 [2024-07-15 18:10:25.834600] bdev_nvme.c:6912:discovery_log_page_cb: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 new subsystem nvme1 00:24:32.279 18:10:25 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # get_bdev_list 00:24:32.279 18:10:25 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # sort 00:24:32.279 18:10:25 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # rpc_cmd -s /tmp/host.sock bdev_get_bdevs 00:24:32.279 18:10:25 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # jq -r '.[].name' 00:24:32.279 18:10:25 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:32.279 18:10:25 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:32.279 18:10:25 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@29 -- # xargs 00:24:32.279 [2024-07-15 18:10:25.971891] bdev_nvme.c:7773:bdev_nvme_readv: *DEBUG*: read 8 blocks with offset 0 00:24:32.279 [2024-07-15 18:10:25.971926] bdev_nvme.c:7773:bdev_nvme_readv: *DEBUG*: read 1 blocks with offset 0 00:24:32.279 [2024-07-15 18:10:25.971946] bdev_nvme.c:7773:bdev_nvme_readv: *DEBUG*: read 64 blocks with offset 0 00:24:32.279 [2024-07-15 18:10:25.971959] bdev_nvme.c:6802:discovery_attach_controller_done: *INFO*: Discovery[10.0.0.2:8009] attach nvme1 done 00:24:32.279 [2024-07-15 18:10:25.971966] bdev_nvme.c:6761:discovery_remove_controllers: *INFO*: Discovery[10.0.0.2:8009] NVM nqn.2016-06.io.spdk:cnode0:10.0.0.2:4420 found again 00:24:32.279 [2024-07-15 18:10:25.977391] bdev_nvme.c:1617:bdev_nvme_disconnected_qpair_cb: *DEBUG*: qpair 0x19aa990 was disconnected and freed. delete nvme_qpair. 00:24:32.279 18:10:25 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:32.538 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@33 -- # [[ nvme1n1 != \n\v\m\e\1\n\1 ]] 00:24:32.538 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@88 -- # trap - SIGINT SIGTERM EXIT 00:24:32.538 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@90 -- # killprocess 705449 00:24:32.538 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@948 -- # '[' -z 705449 ']' 00:24:32.538 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@952 -- # kill -0 705449 00:24:32.538 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@953 -- # uname 00:24:32.538 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:24:32.538 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 705449 00:24:32.538 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:24:32.538 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:24:32.538 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@966 -- # echo 'killing process with pid 705449' 00:24:32.538 killing process with pid 705449 00:24:32.538 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@967 -- # kill 705449 00:24:32.538 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@972 -- # wait 705449 00:24:32.538 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- host/discovery_remove_ifc.sh@91 -- # nvmftestfini 00:24:32.538 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@488 -- # nvmfcleanup 00:24:32.538 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@117 -- # sync 00:24:32.538 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:24:32.538 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@120 -- # set +e 00:24:32.538 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@121 -- # for i in {1..20} 00:24:32.538 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:24:32.538 rmmod nvme_tcp 00:24:32.538 rmmod nvme_fabrics 00:24:32.798 rmmod nvme_keyring 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@124 -- # set -e 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@125 -- # return 0 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@489 -- # '[' -n 705295 ']' 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@490 -- # killprocess 705295 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@948 -- # '[' -z 705295 ']' 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@952 -- # kill -0 705295 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@953 -- # uname 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 705295 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@966 -- # echo 'killing process with pid 705295' 00:24:32.798 killing process with pid 705295 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@967 -- # kill 705295 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@972 -- # wait 705295 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@278 -- # remove_spdk_ns 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:24:32.798 18:10:26 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:35.335 18:10:28 nvmf_tcp.nvmf_discovery_remove_ifc -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:24:35.335 00:24:35.335 real 0m20.675s 00:24:35.335 user 0m26.389s 00:24:35.335 sys 0m5.103s 00:24:35.335 18:10:28 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@1124 -- # xtrace_disable 00:24:35.335 18:10:28 nvmf_tcp.nvmf_discovery_remove_ifc -- common/autotest_common.sh@10 -- # set +x 00:24:35.335 ************************************ 00:24:35.335 END TEST nvmf_discovery_remove_ifc 00:24:35.335 ************************************ 00:24:35.335 18:10:28 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:24:35.335 18:10:28 nvmf_tcp -- nvmf/nvmf.sh@104 -- # run_test nvmf_identify_kernel_target /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify_kernel_nvmf.sh --transport=tcp 00:24:35.335 18:10:28 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:24:35.335 18:10:28 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:24:35.335 18:10:28 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:24:35.335 ************************************ 00:24:35.335 START TEST nvmf_identify_kernel_target 00:24:35.335 ************************************ 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/identify_kernel_nvmf.sh --transport=tcp 00:24:35.335 * Looking for test storage... 00:24:35.335 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@7 -- # uname -s 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- paths/export.sh@5 -- # export PATH 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@47 -- # : 0 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@51 -- # have_pci_nics=0 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@11 -- # nvmftestinit 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@448 -- # prepare_net_devs 00:24:35.335 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@410 -- # local -g is_hw=no 00:24:35.336 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@412 -- # remove_spdk_ns 00:24:35.336 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:35.336 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:24:35.336 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:35.336 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:24:35.336 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:24:35.336 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@285 -- # xtrace_disable 00:24:35.336 18:10:28 nvmf_tcp.nvmf_identify_kernel_target -- common/autotest_common.sh@10 -- # set +x 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@291 -- # pci_devs=() 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@291 -- # local -a pci_devs 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@292 -- # pci_net_devs=() 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@293 -- # pci_drivers=() 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@293 -- # local -A pci_drivers 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@295 -- # net_devs=() 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@295 -- # local -ga net_devs 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@296 -- # e810=() 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@296 -- # local -ga e810 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@297 -- # x722=() 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@297 -- # local -ga x722 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@298 -- # mlx=() 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@298 -- # local -ga mlx 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:24:40.666 Found 0000:86:00.0 (0x8086 - 0x159b) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:24:40.666 Found 0000:86:00.1 (0x8086 - 0x159b) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@390 -- # [[ up == up ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:24:40.666 Found net devices under 0000:86:00.0: cvl_0_0 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@390 -- # [[ up == up ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:24:40.666 Found net devices under 0000:86:00.1: cvl_0_1 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@414 -- # is_hw=yes 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:24:40.666 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:24:40.667 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:24:40.667 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.330 ms 00:24:40.667 00:24:40.667 --- 10.0.0.2 ping statistics --- 00:24:40.667 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:40.667 rtt min/avg/max/mdev = 0.330/0.330/0.330/0.000 ms 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:24:40.667 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:24:40.667 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.229 ms 00:24:40.667 00:24:40.667 --- 10.0.0.1 ping statistics --- 00:24:40.667 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:40.667 rtt min/avg/max/mdev = 0.229/0.229/0.229/0.000 ms 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@422 -- # return 0 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@13 -- # trap 'nvmftestfini || :; clean_kernel_target' EXIT 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@15 -- # get_main_ns_ip 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@741 -- # local ip 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@742 -- # ip_candidates=() 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@742 -- # local -A ip_candidates 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@15 -- # target_ip=10.0.0.1 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@16 -- # configure_kernel_target nqn.2016-06.io.spdk:testnqn 10.0.0.1 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@632 -- # local kernel_name=nqn.2016-06.io.spdk:testnqn kernel_target_ip=10.0.0.1 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@634 -- # nvmet=/sys/kernel/config/nvmet 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@635 -- # kernel_subsystem=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@636 -- # kernel_namespace=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@637 -- # kernel_port=/sys/kernel/config/nvmet/ports/1 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@639 -- # local block nvme 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@641 -- # [[ ! -e /sys/module/nvmet ]] 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@642 -- # modprobe nvmet 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@645 -- # [[ -e /sys/kernel/config/nvmet ]] 00:24:40.667 18:10:33 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@647 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:24:43.201 Waiting for block devices as requested 00:24:43.201 0000:5e:00.0 (8086 0a54): vfio-pci -> nvme 00:24:43.201 0000:00:04.7 (8086 2021): vfio-pci -> ioatdma 00:24:43.201 0000:00:04.6 (8086 2021): vfio-pci -> ioatdma 00:24:43.201 0000:00:04.5 (8086 2021): vfio-pci -> ioatdma 00:24:43.201 0000:00:04.4 (8086 2021): vfio-pci -> ioatdma 00:24:43.201 0000:00:04.3 (8086 2021): vfio-pci -> ioatdma 00:24:43.460 0000:00:04.2 (8086 2021): vfio-pci -> ioatdma 00:24:43.460 0000:00:04.1 (8086 2021): vfio-pci -> ioatdma 00:24:43.460 0000:00:04.0 (8086 2021): vfio-pci -> ioatdma 00:24:43.460 0000:80:04.7 (8086 2021): vfio-pci -> ioatdma 00:24:43.719 0000:80:04.6 (8086 2021): vfio-pci -> ioatdma 00:24:43.719 0000:80:04.5 (8086 2021): vfio-pci -> ioatdma 00:24:43.719 0000:80:04.4 (8086 2021): vfio-pci -> ioatdma 00:24:43.979 0000:80:04.3 (8086 2021): vfio-pci -> ioatdma 00:24:43.979 0000:80:04.2 (8086 2021): vfio-pci -> ioatdma 00:24:43.979 0000:80:04.1 (8086 2021): vfio-pci -> ioatdma 00:24:43.979 0000:80:04.0 (8086 2021): vfio-pci -> ioatdma 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@650 -- # for block in /sys/block/nvme* 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@651 -- # [[ -e /sys/block/nvme0n1 ]] 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@652 -- # is_block_zoned nvme0n1 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- common/autotest_common.sh@1662 -- # local device=nvme0n1 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- common/autotest_common.sh@1664 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- common/autotest_common.sh@1665 -- # [[ none != none ]] 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@653 -- # block_in_use nvme0n1 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- scripts/common.sh@378 -- # local block=nvme0n1 pt 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- scripts/common.sh@387 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:24:44.238 No valid GPT data, bailing 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- scripts/common.sh@391 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- scripts/common.sh@391 -- # pt= 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- scripts/common.sh@392 -- # return 1 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@653 -- # nvme=/dev/nvme0n1 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@656 -- # [[ -b /dev/nvme0n1 ]] 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@658 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@659 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@660 -- # mkdir /sys/kernel/config/nvmet/ports/1 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@665 -- # echo SPDK-nqn.2016-06.io.spdk:testnqn 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@667 -- # echo 1 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@668 -- # echo /dev/nvme0n1 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@669 -- # echo 1 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@671 -- # echo 10.0.0.1 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@672 -- # echo tcp 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@673 -- # echo 4420 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@674 -- # echo ipv4 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@677 -- # ln -s /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn /sys/kernel/config/nvmet/ports/1/subsystems/ 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@680 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -a 10.0.0.1 -t tcp -s 4420 00:24:44.238 00:24:44.238 Discovery Log Number of Records 2, Generation counter 2 00:24:44.238 =====Discovery Log Entry 0====== 00:24:44.238 trtype: tcp 00:24:44.238 adrfam: ipv4 00:24:44.238 subtype: current discovery subsystem 00:24:44.238 treq: not specified, sq flow control disable supported 00:24:44.238 portid: 1 00:24:44.238 trsvcid: 4420 00:24:44.238 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:24:44.238 traddr: 10.0.0.1 00:24:44.238 eflags: none 00:24:44.238 sectype: none 00:24:44.238 =====Discovery Log Entry 1====== 00:24:44.238 trtype: tcp 00:24:44.238 adrfam: ipv4 00:24:44.238 subtype: nvme subsystem 00:24:44.238 treq: not specified, sq flow control disable supported 00:24:44.238 portid: 1 00:24:44.238 trsvcid: 4420 00:24:44.238 subnqn: nqn.2016-06.io.spdk:testnqn 00:24:44.238 traddr: 10.0.0.1 00:24:44.238 eflags: none 00:24:44.238 sectype: none 00:24:44.238 18:10:37 nvmf_tcp.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.1 00:24:44.238 trsvcid:4420 subnqn:nqn.2014-08.org.nvmexpress.discovery' 00:24:44.238 EAL: No free 2048 kB hugepages reported on node 1 00:24:44.498 ===================================================== 00:24:44.498 NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2014-08.org.nvmexpress.discovery 00:24:44.498 ===================================================== 00:24:44.498 Controller Capabilities/Features 00:24:44.498 ================================ 00:24:44.498 Vendor ID: 0000 00:24:44.498 Subsystem Vendor ID: 0000 00:24:44.498 Serial Number: f841570f562f326fb72a 00:24:44.498 Model Number: Linux 00:24:44.498 Firmware Version: 6.7.0-68 00:24:44.498 Recommended Arb Burst: 0 00:24:44.498 IEEE OUI Identifier: 00 00 00 00:24:44.498 Multi-path I/O 00:24:44.498 May have multiple subsystem ports: No 00:24:44.498 May have multiple controllers: No 00:24:44.498 Associated with SR-IOV VF: No 00:24:44.498 Max Data Transfer Size: Unlimited 00:24:44.498 Max Number of Namespaces: 0 00:24:44.498 Max Number of I/O Queues: 1024 00:24:44.498 NVMe Specification Version (VS): 1.3 00:24:44.498 NVMe Specification Version (Identify): 1.3 00:24:44.498 Maximum Queue Entries: 1024 00:24:44.498 Contiguous Queues Required: No 00:24:44.498 Arbitration Mechanisms Supported 00:24:44.498 Weighted Round Robin: Not Supported 00:24:44.498 Vendor Specific: Not Supported 00:24:44.498 Reset Timeout: 7500 ms 00:24:44.498 Doorbell Stride: 4 bytes 00:24:44.498 NVM Subsystem Reset: Not Supported 00:24:44.498 Command Sets Supported 00:24:44.498 NVM Command Set: Supported 00:24:44.498 Boot Partition: Not Supported 00:24:44.498 Memory Page Size Minimum: 4096 bytes 00:24:44.498 Memory Page Size Maximum: 4096 bytes 00:24:44.498 Persistent Memory Region: Not Supported 00:24:44.498 Optional Asynchronous Events Supported 00:24:44.498 Namespace Attribute Notices: Not Supported 00:24:44.498 Firmware Activation Notices: Not Supported 00:24:44.498 ANA Change Notices: Not Supported 00:24:44.498 PLE Aggregate Log Change Notices: Not Supported 00:24:44.498 LBA Status Info Alert Notices: Not Supported 00:24:44.498 EGE Aggregate Log Change Notices: Not Supported 00:24:44.498 Normal NVM Subsystem Shutdown event: Not Supported 00:24:44.498 Zone Descriptor Change Notices: Not Supported 00:24:44.498 Discovery Log Change Notices: Supported 00:24:44.498 Controller Attributes 00:24:44.498 128-bit Host Identifier: Not Supported 00:24:44.498 Non-Operational Permissive Mode: Not Supported 00:24:44.498 NVM Sets: Not Supported 00:24:44.498 Read Recovery Levels: Not Supported 00:24:44.498 Endurance Groups: Not Supported 00:24:44.498 Predictable Latency Mode: Not Supported 00:24:44.498 Traffic Based Keep ALive: Not Supported 00:24:44.498 Namespace Granularity: Not Supported 00:24:44.498 SQ Associations: Not Supported 00:24:44.498 UUID List: Not Supported 00:24:44.498 Multi-Domain Subsystem: Not Supported 00:24:44.498 Fixed Capacity Management: Not Supported 00:24:44.498 Variable Capacity Management: Not Supported 00:24:44.498 Delete Endurance Group: Not Supported 00:24:44.498 Delete NVM Set: Not Supported 00:24:44.498 Extended LBA Formats Supported: Not Supported 00:24:44.498 Flexible Data Placement Supported: Not Supported 00:24:44.498 00:24:44.498 Controller Memory Buffer Support 00:24:44.498 ================================ 00:24:44.498 Supported: No 00:24:44.498 00:24:44.498 Persistent Memory Region Support 00:24:44.498 ================================ 00:24:44.498 Supported: No 00:24:44.498 00:24:44.498 Admin Command Set Attributes 00:24:44.498 ============================ 00:24:44.498 Security Send/Receive: Not Supported 00:24:44.498 Format NVM: Not Supported 00:24:44.498 Firmware Activate/Download: Not Supported 00:24:44.498 Namespace Management: Not Supported 00:24:44.498 Device Self-Test: Not Supported 00:24:44.498 Directives: Not Supported 00:24:44.498 NVMe-MI: Not Supported 00:24:44.498 Virtualization Management: Not Supported 00:24:44.498 Doorbell Buffer Config: Not Supported 00:24:44.498 Get LBA Status Capability: Not Supported 00:24:44.498 Command & Feature Lockdown Capability: Not Supported 00:24:44.498 Abort Command Limit: 1 00:24:44.498 Async Event Request Limit: 1 00:24:44.498 Number of Firmware Slots: N/A 00:24:44.498 Firmware Slot 1 Read-Only: N/A 00:24:44.498 Firmware Activation Without Reset: N/A 00:24:44.498 Multiple Update Detection Support: N/A 00:24:44.498 Firmware Update Granularity: No Information Provided 00:24:44.498 Per-Namespace SMART Log: No 00:24:44.498 Asymmetric Namespace Access Log Page: Not Supported 00:24:44.498 Subsystem NQN: nqn.2014-08.org.nvmexpress.discovery 00:24:44.498 Command Effects Log Page: Not Supported 00:24:44.498 Get Log Page Extended Data: Supported 00:24:44.498 Telemetry Log Pages: Not Supported 00:24:44.498 Persistent Event Log Pages: Not Supported 00:24:44.498 Supported Log Pages Log Page: May Support 00:24:44.498 Commands Supported & Effects Log Page: Not Supported 00:24:44.498 Feature Identifiers & Effects Log Page:May Support 00:24:44.498 NVMe-MI Commands & Effects Log Page: May Support 00:24:44.498 Data Area 4 for Telemetry Log: Not Supported 00:24:44.498 Error Log Page Entries Supported: 1 00:24:44.498 Keep Alive: Not Supported 00:24:44.498 00:24:44.498 NVM Command Set Attributes 00:24:44.498 ========================== 00:24:44.498 Submission Queue Entry Size 00:24:44.498 Max: 1 00:24:44.498 Min: 1 00:24:44.498 Completion Queue Entry Size 00:24:44.498 Max: 1 00:24:44.498 Min: 1 00:24:44.498 Number of Namespaces: 0 00:24:44.498 Compare Command: Not Supported 00:24:44.498 Write Uncorrectable Command: Not Supported 00:24:44.499 Dataset Management Command: Not Supported 00:24:44.499 Write Zeroes Command: Not Supported 00:24:44.499 Set Features Save Field: Not Supported 00:24:44.499 Reservations: Not Supported 00:24:44.499 Timestamp: Not Supported 00:24:44.499 Copy: Not Supported 00:24:44.499 Volatile Write Cache: Not Present 00:24:44.499 Atomic Write Unit (Normal): 1 00:24:44.499 Atomic Write Unit (PFail): 1 00:24:44.499 Atomic Compare & Write Unit: 1 00:24:44.499 Fused Compare & Write: Not Supported 00:24:44.499 Scatter-Gather List 00:24:44.499 SGL Command Set: Supported 00:24:44.499 SGL Keyed: Not Supported 00:24:44.499 SGL Bit Bucket Descriptor: Not Supported 00:24:44.499 SGL Metadata Pointer: Not Supported 00:24:44.499 Oversized SGL: Not Supported 00:24:44.499 SGL Metadata Address: Not Supported 00:24:44.499 SGL Offset: Supported 00:24:44.499 Transport SGL Data Block: Not Supported 00:24:44.499 Replay Protected Memory Block: Not Supported 00:24:44.499 00:24:44.499 Firmware Slot Information 00:24:44.499 ========================= 00:24:44.499 Active slot: 0 00:24:44.499 00:24:44.499 00:24:44.499 Error Log 00:24:44.499 ========= 00:24:44.499 00:24:44.499 Active Namespaces 00:24:44.499 ================= 00:24:44.499 Discovery Log Page 00:24:44.499 ================== 00:24:44.499 Generation Counter: 2 00:24:44.499 Number of Records: 2 00:24:44.499 Record Format: 0 00:24:44.499 00:24:44.499 Discovery Log Entry 0 00:24:44.499 ---------------------- 00:24:44.499 Transport Type: 3 (TCP) 00:24:44.499 Address Family: 1 (IPv4) 00:24:44.499 Subsystem Type: 3 (Current Discovery Subsystem) 00:24:44.499 Entry Flags: 00:24:44.499 Duplicate Returned Information: 0 00:24:44.499 Explicit Persistent Connection Support for Discovery: 0 00:24:44.499 Transport Requirements: 00:24:44.499 Secure Channel: Not Specified 00:24:44.499 Port ID: 1 (0x0001) 00:24:44.499 Controller ID: 65535 (0xffff) 00:24:44.499 Admin Max SQ Size: 32 00:24:44.499 Transport Service Identifier: 4420 00:24:44.499 NVM Subsystem Qualified Name: nqn.2014-08.org.nvmexpress.discovery 00:24:44.499 Transport Address: 10.0.0.1 00:24:44.499 Discovery Log Entry 1 00:24:44.499 ---------------------- 00:24:44.499 Transport Type: 3 (TCP) 00:24:44.499 Address Family: 1 (IPv4) 00:24:44.499 Subsystem Type: 2 (NVM Subsystem) 00:24:44.499 Entry Flags: 00:24:44.499 Duplicate Returned Information: 0 00:24:44.499 Explicit Persistent Connection Support for Discovery: 0 00:24:44.499 Transport Requirements: 00:24:44.499 Secure Channel: Not Specified 00:24:44.499 Port ID: 1 (0x0001) 00:24:44.499 Controller ID: 65535 (0xffff) 00:24:44.499 Admin Max SQ Size: 32 00:24:44.499 Transport Service Identifier: 4420 00:24:44.499 NVM Subsystem Qualified Name: nqn.2016-06.io.spdk:testnqn 00:24:44.499 Transport Address: 10.0.0.1 00:24:44.499 18:10:38 nvmf_tcp.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:24:44.499 EAL: No free 2048 kB hugepages reported on node 1 00:24:44.499 get_feature(0x01) failed 00:24:44.499 get_feature(0x02) failed 00:24:44.499 get_feature(0x04) failed 00:24:44.499 ===================================================== 00:24:44.499 NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:24:44.499 ===================================================== 00:24:44.499 Controller Capabilities/Features 00:24:44.499 ================================ 00:24:44.499 Vendor ID: 0000 00:24:44.499 Subsystem Vendor ID: 0000 00:24:44.499 Serial Number: 580e4416d8408da6deed 00:24:44.499 Model Number: SPDK-nqn.2016-06.io.spdk:testnqn 00:24:44.499 Firmware Version: 6.7.0-68 00:24:44.499 Recommended Arb Burst: 6 00:24:44.499 IEEE OUI Identifier: 00 00 00 00:24:44.499 Multi-path I/O 00:24:44.499 May have multiple subsystem ports: Yes 00:24:44.499 May have multiple controllers: Yes 00:24:44.499 Associated with SR-IOV VF: No 00:24:44.499 Max Data Transfer Size: Unlimited 00:24:44.499 Max Number of Namespaces: 1024 00:24:44.499 Max Number of I/O Queues: 128 00:24:44.499 NVMe Specification Version (VS): 1.3 00:24:44.499 NVMe Specification Version (Identify): 1.3 00:24:44.499 Maximum Queue Entries: 1024 00:24:44.499 Contiguous Queues Required: No 00:24:44.499 Arbitration Mechanisms Supported 00:24:44.499 Weighted Round Robin: Not Supported 00:24:44.499 Vendor Specific: Not Supported 00:24:44.499 Reset Timeout: 7500 ms 00:24:44.499 Doorbell Stride: 4 bytes 00:24:44.499 NVM Subsystem Reset: Not Supported 00:24:44.499 Command Sets Supported 00:24:44.499 NVM Command Set: Supported 00:24:44.499 Boot Partition: Not Supported 00:24:44.499 Memory Page Size Minimum: 4096 bytes 00:24:44.499 Memory Page Size Maximum: 4096 bytes 00:24:44.499 Persistent Memory Region: Not Supported 00:24:44.499 Optional Asynchronous Events Supported 00:24:44.499 Namespace Attribute Notices: Supported 00:24:44.499 Firmware Activation Notices: Not Supported 00:24:44.499 ANA Change Notices: Supported 00:24:44.499 PLE Aggregate Log Change Notices: Not Supported 00:24:44.499 LBA Status Info Alert Notices: Not Supported 00:24:44.499 EGE Aggregate Log Change Notices: Not Supported 00:24:44.499 Normal NVM Subsystem Shutdown event: Not Supported 00:24:44.499 Zone Descriptor Change Notices: Not Supported 00:24:44.499 Discovery Log Change Notices: Not Supported 00:24:44.499 Controller Attributes 00:24:44.499 128-bit Host Identifier: Supported 00:24:44.499 Non-Operational Permissive Mode: Not Supported 00:24:44.499 NVM Sets: Not Supported 00:24:44.499 Read Recovery Levels: Not Supported 00:24:44.499 Endurance Groups: Not Supported 00:24:44.499 Predictable Latency Mode: Not Supported 00:24:44.499 Traffic Based Keep ALive: Supported 00:24:44.499 Namespace Granularity: Not Supported 00:24:44.499 SQ Associations: Not Supported 00:24:44.499 UUID List: Not Supported 00:24:44.499 Multi-Domain Subsystem: Not Supported 00:24:44.499 Fixed Capacity Management: Not Supported 00:24:44.499 Variable Capacity Management: Not Supported 00:24:44.499 Delete Endurance Group: Not Supported 00:24:44.499 Delete NVM Set: Not Supported 00:24:44.499 Extended LBA Formats Supported: Not Supported 00:24:44.499 Flexible Data Placement Supported: Not Supported 00:24:44.499 00:24:44.499 Controller Memory Buffer Support 00:24:44.499 ================================ 00:24:44.499 Supported: No 00:24:44.499 00:24:44.499 Persistent Memory Region Support 00:24:44.499 ================================ 00:24:44.499 Supported: No 00:24:44.499 00:24:44.499 Admin Command Set Attributes 00:24:44.499 ============================ 00:24:44.499 Security Send/Receive: Not Supported 00:24:44.499 Format NVM: Not Supported 00:24:44.499 Firmware Activate/Download: Not Supported 00:24:44.499 Namespace Management: Not Supported 00:24:44.499 Device Self-Test: Not Supported 00:24:44.499 Directives: Not Supported 00:24:44.499 NVMe-MI: Not Supported 00:24:44.499 Virtualization Management: Not Supported 00:24:44.499 Doorbell Buffer Config: Not Supported 00:24:44.499 Get LBA Status Capability: Not Supported 00:24:44.499 Command & Feature Lockdown Capability: Not Supported 00:24:44.499 Abort Command Limit: 4 00:24:44.499 Async Event Request Limit: 4 00:24:44.499 Number of Firmware Slots: N/A 00:24:44.499 Firmware Slot 1 Read-Only: N/A 00:24:44.499 Firmware Activation Without Reset: N/A 00:24:44.499 Multiple Update Detection Support: N/A 00:24:44.499 Firmware Update Granularity: No Information Provided 00:24:44.499 Per-Namespace SMART Log: Yes 00:24:44.499 Asymmetric Namespace Access Log Page: Supported 00:24:44.499 ANA Transition Time : 10 sec 00:24:44.499 00:24:44.499 Asymmetric Namespace Access Capabilities 00:24:44.499 ANA Optimized State : Supported 00:24:44.499 ANA Non-Optimized State : Supported 00:24:44.499 ANA Inaccessible State : Supported 00:24:44.499 ANA Persistent Loss State : Supported 00:24:44.499 ANA Change State : Supported 00:24:44.499 ANAGRPID is not changed : No 00:24:44.499 Non-Zero ANAGRPID for NS Mgmt Cmd : Not Supported 00:24:44.499 00:24:44.499 ANA Group Identifier Maximum : 128 00:24:44.499 Number of ANA Group Identifiers : 128 00:24:44.499 Max Number of Allowed Namespaces : 1024 00:24:44.499 Subsystem NQN: nqn.2016-06.io.spdk:testnqn 00:24:44.499 Command Effects Log Page: Supported 00:24:44.499 Get Log Page Extended Data: Supported 00:24:44.499 Telemetry Log Pages: Not Supported 00:24:44.499 Persistent Event Log Pages: Not Supported 00:24:44.499 Supported Log Pages Log Page: May Support 00:24:44.499 Commands Supported & Effects Log Page: Not Supported 00:24:44.499 Feature Identifiers & Effects Log Page:May Support 00:24:44.499 NVMe-MI Commands & Effects Log Page: May Support 00:24:44.499 Data Area 4 for Telemetry Log: Not Supported 00:24:44.499 Error Log Page Entries Supported: 128 00:24:44.499 Keep Alive: Supported 00:24:44.499 Keep Alive Granularity: 1000 ms 00:24:44.499 00:24:44.499 NVM Command Set Attributes 00:24:44.499 ========================== 00:24:44.499 Submission Queue Entry Size 00:24:44.499 Max: 64 00:24:44.499 Min: 64 00:24:44.499 Completion Queue Entry Size 00:24:44.499 Max: 16 00:24:44.499 Min: 16 00:24:44.499 Number of Namespaces: 1024 00:24:44.499 Compare Command: Not Supported 00:24:44.499 Write Uncorrectable Command: Not Supported 00:24:44.499 Dataset Management Command: Supported 00:24:44.499 Write Zeroes Command: Supported 00:24:44.500 Set Features Save Field: Not Supported 00:24:44.500 Reservations: Not Supported 00:24:44.500 Timestamp: Not Supported 00:24:44.500 Copy: Not Supported 00:24:44.500 Volatile Write Cache: Present 00:24:44.500 Atomic Write Unit (Normal): 1 00:24:44.500 Atomic Write Unit (PFail): 1 00:24:44.500 Atomic Compare & Write Unit: 1 00:24:44.500 Fused Compare & Write: Not Supported 00:24:44.500 Scatter-Gather List 00:24:44.500 SGL Command Set: Supported 00:24:44.500 SGL Keyed: Not Supported 00:24:44.500 SGL Bit Bucket Descriptor: Not Supported 00:24:44.500 SGL Metadata Pointer: Not Supported 00:24:44.500 Oversized SGL: Not Supported 00:24:44.500 SGL Metadata Address: Not Supported 00:24:44.500 SGL Offset: Supported 00:24:44.500 Transport SGL Data Block: Not Supported 00:24:44.500 Replay Protected Memory Block: Not Supported 00:24:44.500 00:24:44.500 Firmware Slot Information 00:24:44.500 ========================= 00:24:44.500 Active slot: 0 00:24:44.500 00:24:44.500 Asymmetric Namespace Access 00:24:44.500 =========================== 00:24:44.500 Change Count : 0 00:24:44.500 Number of ANA Group Descriptors : 1 00:24:44.500 ANA Group Descriptor : 0 00:24:44.500 ANA Group ID : 1 00:24:44.500 Number of NSID Values : 1 00:24:44.500 Change Count : 0 00:24:44.500 ANA State : 1 00:24:44.500 Namespace Identifier : 1 00:24:44.500 00:24:44.500 Commands Supported and Effects 00:24:44.500 ============================== 00:24:44.500 Admin Commands 00:24:44.500 -------------- 00:24:44.500 Get Log Page (02h): Supported 00:24:44.500 Identify (06h): Supported 00:24:44.500 Abort (08h): Supported 00:24:44.500 Set Features (09h): Supported 00:24:44.500 Get Features (0Ah): Supported 00:24:44.500 Asynchronous Event Request (0Ch): Supported 00:24:44.500 Keep Alive (18h): Supported 00:24:44.500 I/O Commands 00:24:44.500 ------------ 00:24:44.500 Flush (00h): Supported 00:24:44.500 Write (01h): Supported LBA-Change 00:24:44.500 Read (02h): Supported 00:24:44.500 Write Zeroes (08h): Supported LBA-Change 00:24:44.500 Dataset Management (09h): Supported 00:24:44.500 00:24:44.500 Error Log 00:24:44.500 ========= 00:24:44.500 Entry: 0 00:24:44.500 Error Count: 0x3 00:24:44.500 Submission Queue Id: 0x0 00:24:44.500 Command Id: 0x5 00:24:44.500 Phase Bit: 0 00:24:44.500 Status Code: 0x2 00:24:44.500 Status Code Type: 0x0 00:24:44.500 Do Not Retry: 1 00:24:44.500 Error Location: 0x28 00:24:44.500 LBA: 0x0 00:24:44.500 Namespace: 0x0 00:24:44.500 Vendor Log Page: 0x0 00:24:44.500 ----------- 00:24:44.500 Entry: 1 00:24:44.500 Error Count: 0x2 00:24:44.500 Submission Queue Id: 0x0 00:24:44.500 Command Id: 0x5 00:24:44.500 Phase Bit: 0 00:24:44.500 Status Code: 0x2 00:24:44.500 Status Code Type: 0x0 00:24:44.500 Do Not Retry: 1 00:24:44.500 Error Location: 0x28 00:24:44.500 LBA: 0x0 00:24:44.500 Namespace: 0x0 00:24:44.500 Vendor Log Page: 0x0 00:24:44.500 ----------- 00:24:44.500 Entry: 2 00:24:44.500 Error Count: 0x1 00:24:44.500 Submission Queue Id: 0x0 00:24:44.500 Command Id: 0x4 00:24:44.500 Phase Bit: 0 00:24:44.500 Status Code: 0x2 00:24:44.500 Status Code Type: 0x0 00:24:44.500 Do Not Retry: 1 00:24:44.500 Error Location: 0x28 00:24:44.500 LBA: 0x0 00:24:44.500 Namespace: 0x0 00:24:44.500 Vendor Log Page: 0x0 00:24:44.500 00:24:44.500 Number of Queues 00:24:44.500 ================ 00:24:44.500 Number of I/O Submission Queues: 128 00:24:44.500 Number of I/O Completion Queues: 128 00:24:44.500 00:24:44.500 ZNS Specific Controller Data 00:24:44.500 ============================ 00:24:44.500 Zone Append Size Limit: 0 00:24:44.500 00:24:44.500 00:24:44.500 Active Namespaces 00:24:44.500 ================= 00:24:44.500 get_feature(0x05) failed 00:24:44.500 Namespace ID:1 00:24:44.500 Command Set Identifier: NVM (00h) 00:24:44.500 Deallocate: Supported 00:24:44.500 Deallocated/Unwritten Error: Not Supported 00:24:44.500 Deallocated Read Value: Unknown 00:24:44.500 Deallocate in Write Zeroes: Not Supported 00:24:44.500 Deallocated Guard Field: 0xFFFF 00:24:44.500 Flush: Supported 00:24:44.500 Reservation: Not Supported 00:24:44.500 Namespace Sharing Capabilities: Multiple Controllers 00:24:44.500 Size (in LBAs): 1953525168 (931GiB) 00:24:44.500 Capacity (in LBAs): 1953525168 (931GiB) 00:24:44.500 Utilization (in LBAs): 1953525168 (931GiB) 00:24:44.500 UUID: 2777f19a-1e46-40ec-9c84-ed91b9239949 00:24:44.500 Thin Provisioning: Not Supported 00:24:44.500 Per-NS Atomic Units: Yes 00:24:44.500 Atomic Boundary Size (Normal): 0 00:24:44.500 Atomic Boundary Size (PFail): 0 00:24:44.500 Atomic Boundary Offset: 0 00:24:44.500 NGUID/EUI64 Never Reused: No 00:24:44.500 ANA group ID: 1 00:24:44.500 Namespace Write Protected: No 00:24:44.500 Number of LBA Formats: 1 00:24:44.500 Current LBA Format: LBA Format #00 00:24:44.500 LBA Format #00: Data Size: 512 Metadata Size: 0 00:24:44.500 00:24:44.500 18:10:38 nvmf_tcp.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@1 -- # nvmftestfini 00:24:44.500 18:10:38 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@488 -- # nvmfcleanup 00:24:44.500 18:10:38 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@117 -- # sync 00:24:44.500 18:10:38 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:24:44.500 18:10:38 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@120 -- # set +e 00:24:44.500 18:10:38 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@121 -- # for i in {1..20} 00:24:44.500 18:10:38 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:24:44.500 rmmod nvme_tcp 00:24:44.500 rmmod nvme_fabrics 00:24:44.500 18:10:38 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:24:44.500 18:10:38 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@124 -- # set -e 00:24:44.500 18:10:38 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@125 -- # return 0 00:24:44.500 18:10:38 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@489 -- # '[' -n '' ']' 00:24:44.500 18:10:38 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:24:44.500 18:10:38 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:24:44.500 18:10:38 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:24:44.500 18:10:38 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:24:44.500 18:10:38 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@278 -- # remove_spdk_ns 00:24:44.500 18:10:38 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:44.500 18:10:38 nvmf_tcp.nvmf_identify_kernel_target -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:24:44.500 18:10:38 nvmf_tcp.nvmf_identify_kernel_target -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:47.034 18:10:40 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:24:47.034 18:10:40 nvmf_tcp.nvmf_identify_kernel_target -- host/identify_kernel_nvmf.sh@1 -- # clean_kernel_target 00:24:47.034 18:10:40 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@684 -- # [[ -e /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn ]] 00:24:47.034 18:10:40 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@686 -- # echo 0 00:24:47.034 18:10:40 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@688 -- # rm -f /sys/kernel/config/nvmet/ports/1/subsystems/nqn.2016-06.io.spdk:testnqn 00:24:47.034 18:10:40 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@689 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:24:47.034 18:10:40 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@690 -- # rmdir /sys/kernel/config/nvmet/ports/1 00:24:47.034 18:10:40 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@691 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:24:47.034 18:10:40 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@693 -- # modules=(/sys/module/nvmet/holders/*) 00:24:47.034 18:10:40 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@695 -- # modprobe -r nvmet_tcp nvmet 00:24:47.034 18:10:40 nvmf_tcp.nvmf_identify_kernel_target -- nvmf/common.sh@698 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:24:49.570 0000:00:04.7 (8086 2021): ioatdma -> vfio-pci 00:24:49.570 0000:00:04.6 (8086 2021): ioatdma -> vfio-pci 00:24:49.570 0000:00:04.5 (8086 2021): ioatdma -> vfio-pci 00:24:49.570 0000:00:04.4 (8086 2021): ioatdma -> vfio-pci 00:24:49.570 0000:00:04.3 (8086 2021): ioatdma -> vfio-pci 00:24:49.570 0000:00:04.2 (8086 2021): ioatdma -> vfio-pci 00:24:49.570 0000:00:04.1 (8086 2021): ioatdma -> vfio-pci 00:24:49.570 0000:00:04.0 (8086 2021): ioatdma -> vfio-pci 00:24:49.570 0000:80:04.7 (8086 2021): ioatdma -> vfio-pci 00:24:49.570 0000:80:04.6 (8086 2021): ioatdma -> vfio-pci 00:24:49.570 0000:80:04.5 (8086 2021): ioatdma -> vfio-pci 00:24:49.570 0000:80:04.4 (8086 2021): ioatdma -> vfio-pci 00:24:49.570 0000:80:04.3 (8086 2021): ioatdma -> vfio-pci 00:24:49.570 0000:80:04.2 (8086 2021): ioatdma -> vfio-pci 00:24:49.570 0000:80:04.1 (8086 2021): ioatdma -> vfio-pci 00:24:49.570 0000:80:04.0 (8086 2021): ioatdma -> vfio-pci 00:24:50.138 0000:5e:00.0 (8086 0a54): nvme -> vfio-pci 00:24:50.138 00:24:50.138 real 0m15.196s 00:24:50.138 user 0m3.665s 00:24:50.138 sys 0m7.887s 00:24:50.138 18:10:43 nvmf_tcp.nvmf_identify_kernel_target -- common/autotest_common.sh@1124 -- # xtrace_disable 00:24:50.138 18:10:43 nvmf_tcp.nvmf_identify_kernel_target -- common/autotest_common.sh@10 -- # set +x 00:24:50.138 ************************************ 00:24:50.138 END TEST nvmf_identify_kernel_target 00:24:50.138 ************************************ 00:24:50.397 18:10:43 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:24:50.397 18:10:43 nvmf_tcp -- nvmf/nvmf.sh@105 -- # run_test nvmf_auth_host /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/auth.sh --transport=tcp 00:24:50.397 18:10:43 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:24:50.397 18:10:43 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:24:50.397 18:10:43 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:24:50.397 ************************************ 00:24:50.397 START TEST nvmf_auth_host 00:24:50.397 ************************************ 00:24:50.397 18:10:43 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/auth.sh --transport=tcp 00:24:50.397 * Looking for test storage... 00:24:50.397 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- host/auth.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@7 -- # uname -s 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:50.397 18:10:44 nvmf_tcp.nvmf_auth_host -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- paths/export.sh@5 -- # export PATH 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@47 -- # : 0 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@51 -- # have_pci_nics=0 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- host/auth.sh@13 -- # digests=("sha256" "sha384" "sha512") 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- host/auth.sh@16 -- # dhgroups=("ffdhe2048" "ffdhe3072" "ffdhe4096" "ffdhe6144" "ffdhe8192") 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- host/auth.sh@17 -- # subnqn=nqn.2024-02.io.spdk:cnode0 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- host/auth.sh@18 -- # hostnqn=nqn.2024-02.io.spdk:host0 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- host/auth.sh@19 -- # nvmet_subsys=/sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- host/auth.sh@20 -- # nvmet_host=/sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- host/auth.sh@21 -- # keys=() 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- host/auth.sh@21 -- # ckeys=() 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- host/auth.sh@68 -- # nvmftestinit 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@448 -- # prepare_net_devs 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@410 -- # local -g is_hw=no 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@412 -- # remove_spdk_ns 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@285 -- # xtrace_disable 00:24:50.398 18:10:44 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@291 -- # pci_devs=() 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@291 -- # local -a pci_devs 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@292 -- # pci_net_devs=() 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@293 -- # pci_drivers=() 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@293 -- # local -A pci_drivers 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@295 -- # net_devs=() 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@295 -- # local -ga net_devs 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@296 -- # e810=() 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@296 -- # local -ga e810 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@297 -- # x722=() 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@297 -- # local -ga x722 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@298 -- # mlx=() 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@298 -- # local -ga mlx 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:24:55.670 Found 0000:86:00.0 (0x8086 - 0x159b) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:24:55.670 Found 0000:86:00.1 (0x8086 - 0x159b) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@390 -- # [[ up == up ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:24:55.670 Found net devices under 0000:86:00.0: cvl_0_0 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@390 -- # [[ up == up ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:24:55.670 Found net devices under 0000:86:00.1: cvl_0_1 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@414 -- # is_hw=yes 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:24:55.670 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:24:55.671 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:24:55.671 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:24:55.671 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:24:55.671 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:24:55.671 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.171 ms 00:24:55.671 00:24:55.671 --- 10.0.0.2 ping statistics --- 00:24:55.671 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:55.671 rtt min/avg/max/mdev = 0.171/0.171/0.171/0.000 ms 00:24:55.671 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:24:55.671 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:24:55.671 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.154 ms 00:24:55.671 00:24:55.671 --- 10.0.0.1 ping statistics --- 00:24:55.671 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:24:55.671 rtt min/avg/max/mdev = 0.154/0.154/0.154/0.000 ms 00:24:55.671 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:24:55.671 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@422 -- # return 0 00:24:55.671 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:24:55.671 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:24:55.671 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:24:55.671 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:24:55.671 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:24:55.671 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:24:55.671 18:10:48 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:24:55.671 18:10:49 nvmf_tcp.nvmf_auth_host -- host/auth.sh@69 -- # nvmfappstart -L nvme_auth 00:24:55.671 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:24:55.671 18:10:49 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@722 -- # xtrace_disable 00:24:55.671 18:10:49 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:24:55.671 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@481 -- # nvmfpid=717195 00:24:55.671 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@482 -- # waitforlisten 717195 00:24:55.671 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -L nvme_auth 00:24:55.671 18:10:49 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@829 -- # '[' -z 717195 ']' 00:24:55.671 18:10:49 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:55.671 18:10:49 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@834 -- # local max_retries=100 00:24:55.671 18:10:49 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:55.671 18:10:49 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@838 -- # xtrace_disable 00:24:55.671 18:10:49 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:24:56.239 18:10:49 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@862 -- # return 0 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@728 -- # xtrace_disable 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- host/auth.sh@70 -- # trap 'cat /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvme-auth.log; cleanup' SIGINT SIGTERM EXIT 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- host/auth.sh@73 -- # gen_dhchap_key null 32 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@723 -- # local digest len file key 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@724 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@724 -- # local -A digests 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@726 -- # digest=null 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@726 -- # len=32 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@727 -- # xxd -p -c0 -l 16 /dev/urandom 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@727 -- # key=e15b97ed883a1fe5ed9f670152ea6510 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@728 -- # mktemp -t spdk.key-null.XXX 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@728 -- # file=/tmp/spdk.key-null.r5r 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@729 -- # format_dhchap_key e15b97ed883a1fe5ed9f670152ea6510 0 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@719 -- # format_key DHHC-1 e15b97ed883a1fe5ed9f670152ea6510 0 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@702 -- # local prefix key digest 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # prefix=DHHC-1 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # key=e15b97ed883a1fe5ed9f670152ea6510 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # digest=0 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@705 -- # python - 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@730 -- # chmod 0600 /tmp/spdk.key-null.r5r 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@732 -- # echo /tmp/spdk.key-null.r5r 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- host/auth.sh@73 -- # keys[0]=/tmp/spdk.key-null.r5r 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- host/auth.sh@73 -- # gen_dhchap_key sha512 64 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@723 -- # local digest len file key 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@724 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@724 -- # local -A digests 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@726 -- # digest=sha512 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@726 -- # len=64 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@727 -- # xxd -p -c0 -l 32 /dev/urandom 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@727 -- # key=f4da44d730636467630d3f4731b3d3f7d12574cfebfef3fe10227b5070343bee 00:24:56.240 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@728 -- # mktemp -t spdk.key-sha512.XXX 00:24:56.499 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@728 -- # file=/tmp/spdk.key-sha512.RZ0 00:24:56.500 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@729 -- # format_dhchap_key f4da44d730636467630d3f4731b3d3f7d12574cfebfef3fe10227b5070343bee 3 00:24:56.500 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@719 -- # format_key DHHC-1 f4da44d730636467630d3f4731b3d3f7d12574cfebfef3fe10227b5070343bee 3 00:24:56.500 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@702 -- # local prefix key digest 00:24:56.500 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # prefix=DHHC-1 00:24:56.500 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # key=f4da44d730636467630d3f4731b3d3f7d12574cfebfef3fe10227b5070343bee 00:24:56.500 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # digest=3 00:24:56.500 18:10:49 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@705 -- # python - 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@730 -- # chmod 0600 /tmp/spdk.key-sha512.RZ0 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@732 -- # echo /tmp/spdk.key-sha512.RZ0 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@73 -- # ckeys[0]=/tmp/spdk.key-sha512.RZ0 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@74 -- # gen_dhchap_key null 48 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@723 -- # local digest len file key 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@724 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@724 -- # local -A digests 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@726 -- # digest=null 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@726 -- # len=48 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@727 -- # xxd -p -c0 -l 24 /dev/urandom 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@727 -- # key=e8024f116c2ca6b6d90542c85c676b842439bc36b79f497e 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@728 -- # mktemp -t spdk.key-null.XXX 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@728 -- # file=/tmp/spdk.key-null.QKj 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@729 -- # format_dhchap_key e8024f116c2ca6b6d90542c85c676b842439bc36b79f497e 0 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@719 -- # format_key DHHC-1 e8024f116c2ca6b6d90542c85c676b842439bc36b79f497e 0 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@702 -- # local prefix key digest 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # prefix=DHHC-1 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # key=e8024f116c2ca6b6d90542c85c676b842439bc36b79f497e 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # digest=0 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@705 -- # python - 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@730 -- # chmod 0600 /tmp/spdk.key-null.QKj 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@732 -- # echo /tmp/spdk.key-null.QKj 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@74 -- # keys[1]=/tmp/spdk.key-null.QKj 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@74 -- # gen_dhchap_key sha384 48 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@723 -- # local digest len file key 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@724 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@724 -- # local -A digests 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@726 -- # digest=sha384 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@726 -- # len=48 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@727 -- # xxd -p -c0 -l 24 /dev/urandom 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@727 -- # key=59d2aad09a87637a628e59ea4341c4e51f54088d679be2e8 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@728 -- # mktemp -t spdk.key-sha384.XXX 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@728 -- # file=/tmp/spdk.key-sha384.dTg 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@729 -- # format_dhchap_key 59d2aad09a87637a628e59ea4341c4e51f54088d679be2e8 2 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@719 -- # format_key DHHC-1 59d2aad09a87637a628e59ea4341c4e51f54088d679be2e8 2 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@702 -- # local prefix key digest 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # prefix=DHHC-1 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # key=59d2aad09a87637a628e59ea4341c4e51f54088d679be2e8 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # digest=2 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@705 -- # python - 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@730 -- # chmod 0600 /tmp/spdk.key-sha384.dTg 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@732 -- # echo /tmp/spdk.key-sha384.dTg 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@74 -- # ckeys[1]=/tmp/spdk.key-sha384.dTg 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@75 -- # gen_dhchap_key sha256 32 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@723 -- # local digest len file key 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@724 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@724 -- # local -A digests 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@726 -- # digest=sha256 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@726 -- # len=32 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@727 -- # xxd -p -c0 -l 16 /dev/urandom 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@727 -- # key=597e4629eaa1b894247760bb966b8f68 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@728 -- # mktemp -t spdk.key-sha256.XXX 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@728 -- # file=/tmp/spdk.key-sha256.JmW 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@729 -- # format_dhchap_key 597e4629eaa1b894247760bb966b8f68 1 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@719 -- # format_key DHHC-1 597e4629eaa1b894247760bb966b8f68 1 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@702 -- # local prefix key digest 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # prefix=DHHC-1 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # key=597e4629eaa1b894247760bb966b8f68 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # digest=1 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@705 -- # python - 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@730 -- # chmod 0600 /tmp/spdk.key-sha256.JmW 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@732 -- # echo /tmp/spdk.key-sha256.JmW 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@75 -- # keys[2]=/tmp/spdk.key-sha256.JmW 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@75 -- # gen_dhchap_key sha256 32 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@723 -- # local digest len file key 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@724 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@724 -- # local -A digests 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@726 -- # digest=sha256 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@726 -- # len=32 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@727 -- # xxd -p -c0 -l 16 /dev/urandom 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@727 -- # key=568e387d655dc58996fed1a705b0efe0 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@728 -- # mktemp -t spdk.key-sha256.XXX 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@728 -- # file=/tmp/spdk.key-sha256.t0v 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@729 -- # format_dhchap_key 568e387d655dc58996fed1a705b0efe0 1 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@719 -- # format_key DHHC-1 568e387d655dc58996fed1a705b0efe0 1 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@702 -- # local prefix key digest 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # prefix=DHHC-1 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # key=568e387d655dc58996fed1a705b0efe0 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # digest=1 00:24:56.500 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@705 -- # python - 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@730 -- # chmod 0600 /tmp/spdk.key-sha256.t0v 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@732 -- # echo /tmp/spdk.key-sha256.t0v 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@75 -- # ckeys[2]=/tmp/spdk.key-sha256.t0v 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@76 -- # gen_dhchap_key sha384 48 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@723 -- # local digest len file key 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@724 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@724 -- # local -A digests 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@726 -- # digest=sha384 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@726 -- # len=48 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@727 -- # xxd -p -c0 -l 24 /dev/urandom 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@727 -- # key=d11359da9f26af91e45dbc1113d452b376413608d01351b8 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@728 -- # mktemp -t spdk.key-sha384.XXX 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@728 -- # file=/tmp/spdk.key-sha384.ygI 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@729 -- # format_dhchap_key d11359da9f26af91e45dbc1113d452b376413608d01351b8 2 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@719 -- # format_key DHHC-1 d11359da9f26af91e45dbc1113d452b376413608d01351b8 2 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@702 -- # local prefix key digest 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # prefix=DHHC-1 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # key=d11359da9f26af91e45dbc1113d452b376413608d01351b8 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # digest=2 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@705 -- # python - 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@730 -- # chmod 0600 /tmp/spdk.key-sha384.ygI 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@732 -- # echo /tmp/spdk.key-sha384.ygI 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@76 -- # keys[3]=/tmp/spdk.key-sha384.ygI 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@76 -- # gen_dhchap_key null 32 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@723 -- # local digest len file key 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@724 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@724 -- # local -A digests 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@726 -- # digest=null 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@726 -- # len=32 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@727 -- # xxd -p -c0 -l 16 /dev/urandom 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@727 -- # key=a3412e3bf730c86342f12a1ec1cfedb5 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@728 -- # mktemp -t spdk.key-null.XXX 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@728 -- # file=/tmp/spdk.key-null.qDQ 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@729 -- # format_dhchap_key a3412e3bf730c86342f12a1ec1cfedb5 0 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@719 -- # format_key DHHC-1 a3412e3bf730c86342f12a1ec1cfedb5 0 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@702 -- # local prefix key digest 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # prefix=DHHC-1 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # key=a3412e3bf730c86342f12a1ec1cfedb5 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # digest=0 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@705 -- # python - 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@730 -- # chmod 0600 /tmp/spdk.key-null.qDQ 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@732 -- # echo /tmp/spdk.key-null.qDQ 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@76 -- # ckeys[3]=/tmp/spdk.key-null.qDQ 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@77 -- # gen_dhchap_key sha512 64 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@723 -- # local digest len file key 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@724 -- # digests=(['null']='0' ['sha256']='1' ['sha384']='2' ['sha512']='3') 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@724 -- # local -A digests 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@726 -- # digest=sha512 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@726 -- # len=64 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@727 -- # xxd -p -c0 -l 32 /dev/urandom 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@727 -- # key=fa407d4a5fcfd56b9f60ae8085762f43f34b7b58ed4703ad5e0caf2f6962cc24 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@728 -- # mktemp -t spdk.key-sha512.XXX 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@728 -- # file=/tmp/spdk.key-sha512.vIO 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@729 -- # format_dhchap_key fa407d4a5fcfd56b9f60ae8085762f43f34b7b58ed4703ad5e0caf2f6962cc24 3 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@719 -- # format_key DHHC-1 fa407d4a5fcfd56b9f60ae8085762f43f34b7b58ed4703ad5e0caf2f6962cc24 3 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@702 -- # local prefix key digest 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # prefix=DHHC-1 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # key=fa407d4a5fcfd56b9f60ae8085762f43f34b7b58ed4703ad5e0caf2f6962cc24 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@704 -- # digest=3 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@705 -- # python - 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@730 -- # chmod 0600 /tmp/spdk.key-sha512.vIO 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@732 -- # echo /tmp/spdk.key-sha512.vIO 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@77 -- # keys[4]=/tmp/spdk.key-sha512.vIO 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@77 -- # ckeys[4]= 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@79 -- # waitforlisten 717195 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@829 -- # '[' -z 717195 ']' 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@834 -- # local max_retries=100 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:24:56.760 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@838 -- # xtrace_disable 00:24:56.760 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@862 -- # return 0 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key0 /tmp/spdk.key-null.r5r 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n /tmp/spdk.key-sha512.RZ0 ]] 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@82 -- # rpc_cmd keyring_file_add_key ckey0 /tmp/spdk.key-sha512.RZ0 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key1 /tmp/spdk.key-null.QKj 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n /tmp/spdk.key-sha384.dTg ]] 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@82 -- # rpc_cmd keyring_file_add_key ckey1 /tmp/spdk.key-sha384.dTg 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key2 /tmp/spdk.key-sha256.JmW 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n /tmp/spdk.key-sha256.t0v ]] 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@82 -- # rpc_cmd keyring_file_add_key ckey2 /tmp/spdk.key-sha256.t0v 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key3 /tmp/spdk.key-sha384.ygI 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n /tmp/spdk.key-null.qDQ ]] 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@82 -- # rpc_cmd keyring_file_add_key ckey3 /tmp/spdk.key-null.qDQ 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@80 -- # for i in "${!keys[@]}" 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@81 -- # rpc_cmd keyring_file_add_key key4 /tmp/spdk.key-sha512.vIO 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@82 -- # [[ -n '' ]] 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@85 -- # nvmet_auth_init 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@35 -- # get_main_ns_ip 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- host/auth.sh@35 -- # configure_kernel_target nqn.2024-02.io.spdk:cnode0 10.0.0.1 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@632 -- # local kernel_name=nqn.2024-02.io.spdk:cnode0 kernel_target_ip=10.0.0.1 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@634 -- # nvmet=/sys/kernel/config/nvmet 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@635 -- # kernel_subsystem=/sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@636 -- # kernel_namespace=/sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/namespaces/1 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@637 -- # kernel_port=/sys/kernel/config/nvmet/ports/1 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@639 -- # local block nvme 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@641 -- # [[ ! -e /sys/module/nvmet ]] 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@642 -- # modprobe nvmet 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@645 -- # [[ -e /sys/kernel/config/nvmet ]] 00:24:57.019 18:10:50 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@647 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:24:59.547 Waiting for block devices as requested 00:24:59.547 0000:5e:00.0 (8086 0a54): vfio-pci -> nvme 00:24:59.803 0000:00:04.7 (8086 2021): vfio-pci -> ioatdma 00:24:59.803 0000:00:04.6 (8086 2021): vfio-pci -> ioatdma 00:24:59.803 0000:00:04.5 (8086 2021): vfio-pci -> ioatdma 00:25:00.061 0000:00:04.4 (8086 2021): vfio-pci -> ioatdma 00:25:00.061 0000:00:04.3 (8086 2021): vfio-pci -> ioatdma 00:25:00.061 0000:00:04.2 (8086 2021): vfio-pci -> ioatdma 00:25:00.061 0000:00:04.1 (8086 2021): vfio-pci -> ioatdma 00:25:00.320 0000:00:04.0 (8086 2021): vfio-pci -> ioatdma 00:25:00.320 0000:80:04.7 (8086 2021): vfio-pci -> ioatdma 00:25:00.320 0000:80:04.6 (8086 2021): vfio-pci -> ioatdma 00:25:00.320 0000:80:04.5 (8086 2021): vfio-pci -> ioatdma 00:25:00.587 0000:80:04.4 (8086 2021): vfio-pci -> ioatdma 00:25:00.587 0000:80:04.3 (8086 2021): vfio-pci -> ioatdma 00:25:00.587 0000:80:04.2 (8086 2021): vfio-pci -> ioatdma 00:25:00.864 0000:80:04.1 (8086 2021): vfio-pci -> ioatdma 00:25:00.864 0000:80:04.0 (8086 2021): vfio-pci -> ioatdma 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@650 -- # for block in /sys/block/nvme* 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@651 -- # [[ -e /sys/block/nvme0n1 ]] 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@652 -- # is_block_zoned nvme0n1 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@1662 -- # local device=nvme0n1 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@1664 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@1665 -- # [[ none != none ]] 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@653 -- # block_in_use nvme0n1 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- scripts/common.sh@378 -- # local block=nvme0n1 pt 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- scripts/common.sh@387 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:25:01.434 No valid GPT data, bailing 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- scripts/common.sh@391 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- scripts/common.sh@391 -- # pt= 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- scripts/common.sh@392 -- # return 1 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@653 -- # nvme=/dev/nvme0n1 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@656 -- # [[ -b /dev/nvme0n1 ]] 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@658 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@659 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/namespaces/1 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@660 -- # mkdir /sys/kernel/config/nvmet/ports/1 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@665 -- # echo SPDK-nqn.2024-02.io.spdk:cnode0 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@667 -- # echo 1 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@668 -- # echo /dev/nvme0n1 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@669 -- # echo 1 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@671 -- # echo 10.0.0.1 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@672 -- # echo tcp 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@673 -- # echo 4420 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@674 -- # echo ipv4 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@677 -- # ln -s /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 /sys/kernel/config/nvmet/ports/1/subsystems/ 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@680 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -a 10.0.0.1 -t tcp -s 4420 00:25:01.434 00:25:01.434 Discovery Log Number of Records 2, Generation counter 2 00:25:01.434 =====Discovery Log Entry 0====== 00:25:01.434 trtype: tcp 00:25:01.434 adrfam: ipv4 00:25:01.434 subtype: current discovery subsystem 00:25:01.434 treq: not specified, sq flow control disable supported 00:25:01.434 portid: 1 00:25:01.434 trsvcid: 4420 00:25:01.434 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:25:01.434 traddr: 10.0.0.1 00:25:01.434 eflags: none 00:25:01.434 sectype: none 00:25:01.434 =====Discovery Log Entry 1====== 00:25:01.434 trtype: tcp 00:25:01.434 adrfam: ipv4 00:25:01.434 subtype: nvme subsystem 00:25:01.434 treq: not specified, sq flow control disable supported 00:25:01.434 portid: 1 00:25:01.434 trsvcid: 4420 00:25:01.434 subnqn: nqn.2024-02.io.spdk:cnode0 00:25:01.434 traddr: 10.0.0.1 00:25:01.434 eflags: none 00:25:01.434 sectype: none 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@36 -- # mkdir /sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@37 -- # echo 0 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@38 -- # ln -s /sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/allowed_hosts/nqn.2024-02.io.spdk:host0 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@88 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: ]] 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@93 -- # IFS=, 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@94 -- # printf %s sha256,sha384,sha512 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@93 -- # IFS=, 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@94 -- # printf %s ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@93 -- # connect_authenticate sha256,sha384,sha512 ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 1 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256,sha384,sha512 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256,sha384,sha512 --dhchap-dhgroups ffdhe2048,ffdhe3072,ffdhe4096,ffdhe6144,ffdhe8192 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:01.434 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:01.694 nvme0n1 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@100 -- # for digest in "${digests[@]}" 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 0 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: ]] 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 0 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:01.694 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:01.695 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:01.695 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:01.695 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:01.695 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:01.695 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:01.695 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:01.695 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:01.695 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:01.695 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:01.954 nvme0n1 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:01.954 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: ]] 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 1 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:01.955 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:02.214 nvme0n1 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 2 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: ]] 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 2 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:02.214 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:02.474 nvme0n1 00:25:02.474 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:02.474 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:02.474 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:02.474 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:02.474 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:02.474 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:02.474 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:02.474 18:10:55 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:02.474 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:02.474 18:10:55 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 3 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: ]] 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 3 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:02.474 nvme0n1 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:02.474 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:02.475 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe2048 4 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe2048 4 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:02.735 nvme0n1 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 0 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: ]] 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 0 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:02.735 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:02.995 nvme0n1 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 1 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: ]] 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 1 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:02.995 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:03.255 nvme0n1 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 2 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: ]] 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 2 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:03.255 18:10:56 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:03.514 nvme0n1 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 3 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: ]] 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 3 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:03.514 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:03.773 nvme0n1 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe3072 4 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe3072 4 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe3072 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:03.773 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:04.033 nvme0n1 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 0 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: ]] 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 0 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:04.033 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:04.292 nvme0n1 00:25:04.292 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:04.292 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:04.292 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:04.292 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:04.292 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:04.292 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:04.292 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:04.292 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:04.292 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:04.292 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:04.292 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:04.292 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:04.292 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 1 00:25:04.292 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:04.292 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:04.292 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:25:04.292 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:25:04.292 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:04.292 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: ]] 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 1 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:04.293 18:10:57 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:04.552 nvme0n1 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 2 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: ]] 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 2 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:04.552 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:04.811 nvme0n1 00:25:04.811 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:04.811 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:04.811 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:04.811 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:04.811 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:04.811 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:04.812 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:04.812 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:04.812 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:04.812 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 3 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: ]] 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 3 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:05.071 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:05.330 nvme0n1 00:25:05.330 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:05.330 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:05.330 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:05.330 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:05.330 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:05.330 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe4096 4 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe4096 4 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe4096 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:05.331 18:10:58 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:05.590 nvme0n1 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 0 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: ]] 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 0 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:05.590 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:05.591 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:05.849 nvme0n1 00:25:05.849 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:05.849 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:05.849 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:05.849 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:05.849 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:05.849 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:06.107 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 1 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: ]] 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 1 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:06.108 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:06.366 nvme0n1 00:25:06.366 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:06.366 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:06.366 18:10:59 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:06.366 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:06.366 18:10:59 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:06.366 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:06.366 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:06.366 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:06.366 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:06.366 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:06.366 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:06.366 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:06.366 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 2 00:25:06.366 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:06.366 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:06.366 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:25:06.366 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:25:06.366 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:06.366 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:06.366 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:06.366 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: ]] 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 2 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:06.367 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:06.934 nvme0n1 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 3 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: ]] 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 3 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:06.934 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:07.193 nvme0n1 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe6144 4 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe6144 4 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe6144 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:07.193 18:11:00 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:07.762 nvme0n1 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 0 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: ]] 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 0 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:07.762 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:08.329 nvme0n1 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 1 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: ]] 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 1 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:08.329 18:11:01 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:08.330 18:11:01 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:08.330 18:11:01 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:08.330 18:11:01 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:08.330 18:11:01 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:08.330 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:08.330 18:11:01 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:08.897 nvme0n1 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 2 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: ]] 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 2 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:08.897 18:11:02 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:09.463 nvme0n1 00:25:09.463 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:09.463 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:09.463 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:09.463 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:09.463 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 3 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: ]] 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 3 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:09.722 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:09.723 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:09.723 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:09.723 18:11:03 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:09.723 18:11:03 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:09.723 18:11:03 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:09.723 18:11:03 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:09.723 18:11:03 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:09.723 18:11:03 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:09.723 18:11:03 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:09.723 18:11:03 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:09.723 18:11:03 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:09.723 18:11:03 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:09.723 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:25:09.723 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:09.723 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:10.290 nvme0n1 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha256 ffdhe8192 4 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha256 ffdhe8192 4 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha256 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe8192 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:10.290 18:11:03 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:10.859 nvme0n1 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@100 -- # for digest in "${digests[@]}" 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 0 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: ]] 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 0 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:10.859 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:11.119 nvme0n1 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 1 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: ]] 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 1 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:11.119 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:11.378 nvme0n1 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 2 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: ]] 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 2 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:11.378 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:11.379 18:11:04 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:11.379 18:11:04 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:11.379 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:11.379 18:11:04 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:11.379 nvme0n1 00:25:11.379 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:11.637 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:11.637 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:11.637 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:11.637 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:11.637 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:11.637 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:11.637 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 3 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: ]] 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 3 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:11.638 nvme0n1 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:11.638 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe2048 4 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe2048 4 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe2048 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:11.897 nvme0n1 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:11.897 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 0 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: ]] 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 0 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:11.898 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:12.157 nvme0n1 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 1 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: ]] 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 1 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:12.157 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:25:12.158 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:12.158 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:12.158 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:12.158 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:12.158 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:12.158 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:12.158 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:12.158 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:12.158 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:12.158 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:12.158 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:12.158 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:12.158 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:12.158 18:11:05 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:12.158 18:11:05 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:12.158 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:12.158 18:11:05 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:12.417 nvme0n1 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 2 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: ]] 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 2 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:12.417 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:12.676 nvme0n1 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 3 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: ]] 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 3 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:12.676 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:12.935 nvme0n1 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe3072 4 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe3072 4 00:25:12.935 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe3072 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:12.936 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:13.195 nvme0n1 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 0 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: ]] 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 0 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:13.195 18:11:06 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:13.454 nvme0n1 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 1 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: ]] 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 1 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:13.454 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:13.455 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:13.455 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:13.455 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:13.455 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:13.455 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:13.455 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:13.455 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:13.455 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:13.455 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:13.455 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:13.455 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:13.713 nvme0n1 00:25:13.713 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:13.713 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:13.713 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:13.713 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:13.713 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:13.713 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:13.713 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:13.714 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:13.714 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:13.714 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 2 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: ]] 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 2 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:13.973 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:14.232 nvme0n1 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 3 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: ]] 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 3 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:14.232 18:11:07 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:14.491 nvme0n1 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe4096 4 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe4096 4 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe4096 00:25:14.491 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:14.515 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:14.515 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:14.515 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:14.515 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:14.515 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:14.515 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:14.515 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:14.515 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:14.515 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:14.515 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:14.515 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:14.515 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:14.515 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:14.515 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:25:14.515 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:14.515 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:14.831 nvme0n1 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 0 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: ]] 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 0 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:14.831 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:15.089 nvme0n1 00:25:15.089 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:15.089 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:15.089 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:15.089 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:15.089 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:15.089 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:15.089 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:15.089 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:15.089 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:15.089 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 1 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: ]] 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 1 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:15.348 18:11:08 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:15.606 nvme0n1 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 2 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: ]] 00:25:15.606 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 2 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:15.607 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:16.174 nvme0n1 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 3 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: ]] 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 3 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:16.174 18:11:09 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:16.432 nvme0n1 00:25:16.433 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:16.433 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:16.433 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:16.433 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:16.433 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:16.433 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:16.433 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe6144 4 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe6144 4 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe6144 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:16.692 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:16.951 nvme0n1 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 0 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: ]] 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 0 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:16.951 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:16.952 18:11:10 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:17.519 nvme0n1 00:25:17.519 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:17.519 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:17.519 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:17.519 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:17.519 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:17.519 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:17.519 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:17.519 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:17.519 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:17.519 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:17.777 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:17.777 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:17.777 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 1 00:25:17.777 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:17.777 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:17.777 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: ]] 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 1 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:17.778 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:18.346 nvme0n1 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 2 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: ]] 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 2 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:18.346 18:11:11 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:18.914 nvme0n1 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 3 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: ]] 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 3 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:18.914 18:11:12 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:19.481 nvme0n1 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha384 ffdhe8192 4 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha384 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha384)' 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha384 ffdhe8192 4 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha384 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha384 --dhchap-dhgroups ffdhe8192 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:19.481 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:20.059 nvme0n1 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@100 -- # for digest in "${digests[@]}" 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 0 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: ]] 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 0 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:20.059 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:20.319 nvme0n1 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 1 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: ]] 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:20.319 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 1 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:20.320 18:11:13 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:20.579 nvme0n1 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 2 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: ]] 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 2 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:20.579 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:20.580 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:20.580 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:20.580 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:20.580 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:20.580 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:20.580 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:20.580 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:20.580 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:20.580 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:20.580 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:20.580 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:20.580 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:20.839 nvme0n1 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 3 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: ]] 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 3 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:20.839 nvme0n1 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:20.839 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe2048 4 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe2048 4 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe2048 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe2048 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:21.099 nvme0n1 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 0 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: ]] 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 0 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:21.099 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:21.359 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:21.359 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:21.359 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:21.359 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:21.359 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:21.359 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:21.359 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:21.359 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:21.359 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:21.359 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:21.359 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:21.359 18:11:14 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:21.359 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:21.359 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:21.359 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:21.359 nvme0n1 00:25:21.359 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:21.359 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:21.359 18:11:14 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:21.359 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:21.359 18:11:14 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 1 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: ]] 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 1 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:21.359 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:21.617 nvme0n1 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 2 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: ]] 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 2 00:25:21.617 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:21.618 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:21.876 nvme0n1 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 3 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: ]] 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 3 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:21.876 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:22.134 nvme0n1 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe3072 4 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe3072 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe3072 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe3072 4 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe3072 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:25:22.134 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:22.135 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe3072 00:25:22.135 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:22.135 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:22.135 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:22.135 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:22.135 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:22.135 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:22.135 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:22.135 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:22.135 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:22.135 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:22.135 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:22.135 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:22.135 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:22.135 18:11:15 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:22.135 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:25:22.135 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:22.135 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:22.394 nvme0n1 00:25:22.394 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:22.394 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:22.394 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:22.394 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:22.394 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:22.394 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:22.394 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:22.394 18:11:15 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:22.394 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:22.394 18:11:15 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 0 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: ]] 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 0 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:22.394 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:25:22.395 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:22.395 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:22.395 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:22.395 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:22.395 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:22.395 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:22.395 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:22.395 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:22.395 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:22.395 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:22.395 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:22.395 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:22.395 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:22.395 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:22.395 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:22.395 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:22.395 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:22.654 nvme0n1 00:25:22.654 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:22.654 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:22.654 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:22.654 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:22.654 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:22.654 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:22.654 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:22.654 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:22.654 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:22.654 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:22.654 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:22.654 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 1 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: ]] 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 1 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:22.655 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:22.914 nvme0n1 00:25:22.914 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:22.914 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:22.914 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:22.914 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:22.914 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:22.914 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:22.914 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:22.914 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:22.914 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:22.914 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 2 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: ]] 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 2 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:23.173 nvme0n1 00:25:23.173 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 3 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: ]] 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 3 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:25:23.432 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:25:23.433 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:23.433 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:25:23.433 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:23.433 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:23.433 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:23.433 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:23.433 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:23.433 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:23.433 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:23.433 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:23.433 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:23.433 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:23.433 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:23.433 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:23.433 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:23.433 18:11:16 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:23.433 18:11:16 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:25:23.433 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:23.433 18:11:16 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:23.692 nvme0n1 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe4096 4 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe4096 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe4096 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe4096 4 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe4096 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe4096 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:23.692 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:23.951 nvme0n1 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 0 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: ]] 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 0 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:23.951 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:24.518 nvme0n1 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 1 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: ]] 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 1 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:24.518 18:11:17 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:24.518 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:24.518 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:24.518 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:24.518 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:24.518 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:24.518 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:24.518 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:24.518 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:24.518 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:24.518 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:24.518 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:24.518 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:24.776 nvme0n1 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 2 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: ]] 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 2 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:24.777 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:25.344 nvme0n1 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 3 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:25.344 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: ]] 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 3 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:25.345 18:11:18 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:25.602 nvme0n1 00:25:25.602 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:25.602 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:25.602 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:25.602 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:25.602 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:25.602 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:25.602 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:25.602 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:25.602 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:25.602 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:25.602 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:25.602 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe6144 4 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe6144 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe6144 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe6144 4 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe6144 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe6144 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:25.603 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:25.861 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:25.861 18:11:19 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:25.861 18:11:19 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:25.861 18:11:19 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:25.861 18:11:19 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:25.861 18:11:19 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:25.861 18:11:19 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:25.861 18:11:19 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:25.861 18:11:19 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:25.861 18:11:19 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:25.861 18:11:19 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:25.861 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:25:25.861 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:25.861 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:26.118 nvme0n1 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@101 -- # for dhgroup in "${dhgroups[@]}" 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 0 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=0 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTE1Yjk3ZWQ4ODNhMWZlNWVkOWY2NzAxNTJlYTY1MTCn1Gjt: 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: ]] 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:03:ZjRkYTQ0ZDczMDYzNjQ2NzYzMGQzZjQ3MzFiM2QzZjdkMTI1NzRjZmViZmVmM2ZlMTAyMjdiNTA3MDM0M2JlZYzd2KU=: 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 0 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=0 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:26.118 18:11:19 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:26.119 18:11:19 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:26.119 18:11:19 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:26.119 18:11:19 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:26.119 18:11:19 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:26.119 18:11:19 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:26.119 18:11:19 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:26.119 18:11:19 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:26.119 18:11:19 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:26.119 18:11:19 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:26.119 18:11:19 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key0 --dhchap-ctrlr-key ckey0 00:25:26.119 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:26.119 18:11:19 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:26.686 nvme0n1 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 1 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:25:26.686 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:26.687 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: ]] 00:25:26.687 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:26.687 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 1 00:25:26.687 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:26.687 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:26.687 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:25:26.687 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=1 00:25:26.687 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:26.687 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:25:26.687 18:11:20 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:26.687 18:11:20 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:26.687 18:11:20 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:26.687 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:26.687 18:11:20 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:26.687 18:11:20 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:26.687 18:11:20 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:26.687 18:11:20 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:26.687 18:11:20 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:26.687 18:11:20 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:26.687 18:11:20 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:26.945 18:11:20 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:26.945 18:11:20 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:26.945 18:11:20 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:26.946 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey1 00:25:26.946 18:11:20 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:26.946 18:11:20 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:27.514 nvme0n1 00:25:27.514 18:11:20 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:27.514 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:27.514 18:11:20 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:27.514 18:11:20 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:27.514 18:11:20 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 2 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=2 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:01:NTk3ZTQ2MjllYWExYjg5NDI0Nzc2MGJiOTY2YjhmNjjoy0lQ: 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: ]] 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:01:NTY4ZTM4N2Q2NTVkYzU4OTk2ZmVkMWE3MDViMGVmZTAlg+i9: 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 2 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=2 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 --dhchap-ctrlr-key ckey2 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:27.514 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:28.082 nvme0n1 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 3 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=3 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:02:ZDExMzU5ZGE5ZjI2YWY5MWU0NWRiYzExMTNkNDUyYjM3NjQxMzYwOGQwMTM1MWI4vmkFTA==: 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: ]] 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:00:YTM0MTJlM2JmNzMwYzg2MzQyZjEyYTFlYzFjZmVkYjWF16uS: 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 3 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=3 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:28.082 18:11:21 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:28.083 18:11:21 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:28.083 18:11:21 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:28.083 18:11:21 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key3 --dhchap-ctrlr-key ckey3 00:25:28.083 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:28.083 18:11:21 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:28.651 nvme0n1 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@102 -- # for keyid in "${!keys[@]}" 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@103 -- # nvmet_auth_set_key sha512 ffdhe8192 4 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha512 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe8192 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=4 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey= 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha512)' 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe8192 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:03:ZmE0MDdkNGE1ZmNmZDU2YjlmNjBhZTgwODU3NjJmNDNmMzRiN2I1OGVkNDcwM2FkNWUwY2FmMmY2OTYyY2MyNCF+ugQ=: 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z '' ]] 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@104 -- # connect_authenticate sha512 ffdhe8192 4 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@55 -- # local digest dhgroup keyid ckey 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # digest=sha512 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # dhgroup=ffdhe8192 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@57 -- # keyid=4 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@58 -- # ckey=(${ckeys[keyid]:+--dhchap-ctrlr-key "ckey${keyid}"}) 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@60 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha512 --dhchap-dhgroups ffdhe8192 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # get_main_ns_ip 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@61 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key4 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:28.651 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:29.254 nvme0n1 00:25:29.254 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:29.254 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # rpc_cmd bdev_nvme_get_controllers 00:25:29.254 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # jq -r '.[].name' 00:25:29.254 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:29.254 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:29.254 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:29.254 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@64 -- # [[ nvme0 == \n\v\m\e\0 ]] 00:25:29.254 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@65 -- # rpc_cmd bdev_nvme_detach_controller nvme0 00:25:29.254 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:29.254 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:29.254 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:29.254 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@110 -- # nvmet_auth_set_key sha256 ffdhe2048 1 00:25:29.254 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@42 -- # local digest dhgroup keyid key ckey 00:25:29.512 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # digest=sha256 00:25:29.512 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # dhgroup=ffdhe2048 00:25:29.512 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@44 -- # keyid=1 00:25:29.512 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@45 -- # key=DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@46 -- # ckey=DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@48 -- # echo 'hmac(sha256)' 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@49 -- # echo ffdhe2048 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@50 -- # echo DHHC-1:00:ZTgwMjRmMTE2YzJjYTZiNmQ5MDU0MmM4NWM2NzZiODQyNDM5YmMzNmI3OWY0OTdlhiczyQ==: 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # [[ -z DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: ]] 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@51 -- # echo DHHC-1:02:NTlkMmFhZDA5YTg3NjM3YTYyOGU1OWVhNDM0MWM0ZTUxZjU0MDg4ZDY3OWJlMmU417vqew==: 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@111 -- # rpc_cmd bdev_nvme_set_options --dhchap-digests sha256 --dhchap-dhgroups ffdhe2048 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@112 -- # get_main_ns_ip 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- host/auth.sh@112 -- # NOT rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@648 -- # local es=0 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@650 -- # valid_exec_arg rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@636 -- # local arg=rpc_cmd 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@640 -- # type -t rpc_cmd 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@651 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:29.513 18:11:22 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:29.513 request: 00:25:29.513 { 00:25:29.513 "name": "nvme0", 00:25:29.513 "trtype": "tcp", 00:25:29.513 "traddr": "10.0.0.1", 00:25:29.513 "adrfam": "ipv4", 00:25:29.513 "trsvcid": "4420", 00:25:29.513 "subnqn": "nqn.2024-02.io.spdk:cnode0", 00:25:29.513 "hostnqn": "nqn.2024-02.io.spdk:host0", 00:25:29.513 "prchk_reftag": false, 00:25:29.513 "prchk_guard": false, 00:25:29.513 "hdgst": false, 00:25:29.513 "ddgst": false, 00:25:29.513 "method": "bdev_nvme_attach_controller", 00:25:29.513 "req_id": 1 00:25:29.513 } 00:25:29.513 Got JSON-RPC error response 00:25:29.513 response: 00:25:29.513 { 00:25:29.513 "code": -5, 00:25:29.513 "message": "Input/output error" 00:25:29.513 } 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 1 == 0 ]] 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@651 -- # es=1 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- host/auth.sh@114 -- # rpc_cmd bdev_nvme_get_controllers 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- host/auth.sh@114 -- # jq length 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- host/auth.sh@114 -- # (( 0 == 0 )) 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- host/auth.sh@117 -- # get_main_ns_ip 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- host/auth.sh@117 -- # NOT rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@648 -- # local es=0 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@650 -- # valid_exec_arg rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@636 -- # local arg=rpc_cmd 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@640 -- # type -t rpc_cmd 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@651 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key2 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:29.513 request: 00:25:29.513 { 00:25:29.513 "name": "nvme0", 00:25:29.513 "trtype": "tcp", 00:25:29.513 "traddr": "10.0.0.1", 00:25:29.513 "adrfam": "ipv4", 00:25:29.513 "trsvcid": "4420", 00:25:29.513 "subnqn": "nqn.2024-02.io.spdk:cnode0", 00:25:29.513 "hostnqn": "nqn.2024-02.io.spdk:host0", 00:25:29.513 "prchk_reftag": false, 00:25:29.513 "prchk_guard": false, 00:25:29.513 "hdgst": false, 00:25:29.513 "ddgst": false, 00:25:29.513 "dhchap_key": "key2", 00:25:29.513 "method": "bdev_nvme_attach_controller", 00:25:29.513 "req_id": 1 00:25:29.513 } 00:25:29.513 Got JSON-RPC error response 00:25:29.513 response: 00:25:29.513 { 00:25:29.513 "code": -5, 00:25:29.513 "message": "Input/output error" 00:25:29.513 } 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 1 == 0 ]] 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@651 -- # es=1 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- host/auth.sh@120 -- # rpc_cmd bdev_nvme_get_controllers 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- host/auth.sh@120 -- # jq length 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- host/auth.sh@120 -- # (( 0 == 0 )) 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- host/auth.sh@123 -- # get_main_ns_ip 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@741 -- # local ip 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # ip_candidates=() 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@742 -- # local -A ip_candidates 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- host/auth.sh@123 -- # NOT rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@648 -- # local es=0 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@650 -- # valid_exec_arg rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@636 -- # local arg=rpc_cmd 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@640 -- # type -t rpc_cmd 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@651 -- # rpc_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -f ipv4 -a 10.0.0.1 -s 4420 -q nqn.2024-02.io.spdk:host0 -n nqn.2024-02.io.spdk:cnode0 --dhchap-key key1 --dhchap-ctrlr-key ckey2 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:29.513 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:29.771 request: 00:25:29.771 { 00:25:29.771 "name": "nvme0", 00:25:29.771 "trtype": "tcp", 00:25:29.771 "traddr": "10.0.0.1", 00:25:29.771 "adrfam": "ipv4", 00:25:29.771 "trsvcid": "4420", 00:25:29.771 "subnqn": "nqn.2024-02.io.spdk:cnode0", 00:25:29.771 "hostnqn": "nqn.2024-02.io.spdk:host0", 00:25:29.771 "prchk_reftag": false, 00:25:29.771 "prchk_guard": false, 00:25:29.771 "hdgst": false, 00:25:29.771 "ddgst": false, 00:25:29.771 "dhchap_key": "key1", 00:25:29.771 "dhchap_ctrlr_key": "ckey2", 00:25:29.771 "method": "bdev_nvme_attach_controller", 00:25:29.771 "req_id": 1 00:25:29.771 } 00:25:29.771 Got JSON-RPC error response 00:25:29.771 response: 00:25:29.771 { 00:25:29.771 "code": -5, 00:25:29.771 "message": "Input/output error" 00:25:29.771 } 00:25:29.771 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@587 -- # [[ 1 == 0 ]] 00:25:29.771 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@651 -- # es=1 00:25:29.771 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:25:29.771 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:25:29.771 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:25:29.771 18:11:23 nvmf_tcp.nvmf_auth_host -- host/auth.sh@127 -- # trap - SIGINT SIGTERM EXIT 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- host/auth.sh@128 -- # cleanup 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- host/auth.sh@24 -- # nvmftestfini 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@488 -- # nvmfcleanup 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@117 -- # sync 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@120 -- # set +e 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@121 -- # for i in {1..20} 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:25:29.772 rmmod nvme_tcp 00:25:29.772 rmmod nvme_fabrics 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@124 -- # set -e 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@125 -- # return 0 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@489 -- # '[' -n 717195 ']' 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@490 -- # killprocess 717195 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@948 -- # '[' -z 717195 ']' 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@952 -- # kill -0 717195 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@953 -- # uname 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 717195 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@966 -- # echo 'killing process with pid 717195' 00:25:29.772 killing process with pid 717195 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@967 -- # kill 717195 00:25:29.772 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@972 -- # wait 717195 00:25:30.030 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:25:30.030 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:25:30.030 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:25:30.030 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:25:30.030 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@278 -- # remove_spdk_ns 00:25:30.030 18:11:23 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:30.030 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:25:30.030 18:11:23 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:31.934 18:11:25 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:25:31.934 18:11:25 nvmf_tcp.nvmf_auth_host -- host/auth.sh@25 -- # rm /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/allowed_hosts/nqn.2024-02.io.spdk:host0 00:25:31.934 18:11:25 nvmf_tcp.nvmf_auth_host -- host/auth.sh@26 -- # rmdir /sys/kernel/config/nvmet/hosts/nqn.2024-02.io.spdk:host0 00:25:31.934 18:11:25 nvmf_tcp.nvmf_auth_host -- host/auth.sh@27 -- # clean_kernel_target 00:25:31.934 18:11:25 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@684 -- # [[ -e /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 ]] 00:25:31.934 18:11:25 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@686 -- # echo 0 00:25:31.934 18:11:25 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@688 -- # rm -f /sys/kernel/config/nvmet/ports/1/subsystems/nqn.2024-02.io.spdk:cnode0 00:25:31.934 18:11:25 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@689 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0/namespaces/1 00:25:31.934 18:11:25 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@690 -- # rmdir /sys/kernel/config/nvmet/ports/1 00:25:31.934 18:11:25 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@691 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2024-02.io.spdk:cnode0 00:25:31.934 18:11:25 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@693 -- # modules=(/sys/module/nvmet/holders/*) 00:25:31.934 18:11:25 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@695 -- # modprobe -r nvmet_tcp nvmet 00:25:31.934 18:11:25 nvmf_tcp.nvmf_auth_host -- nvmf/common.sh@698 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:25:34.469 0000:00:04.7 (8086 2021): ioatdma -> vfio-pci 00:25:34.469 0000:00:04.6 (8086 2021): ioatdma -> vfio-pci 00:25:34.469 0000:00:04.5 (8086 2021): ioatdma -> vfio-pci 00:25:34.469 0000:00:04.4 (8086 2021): ioatdma -> vfio-pci 00:25:34.469 0000:00:04.3 (8086 2021): ioatdma -> vfio-pci 00:25:34.469 0000:00:04.2 (8086 2021): ioatdma -> vfio-pci 00:25:34.469 0000:00:04.1 (8086 2021): ioatdma -> vfio-pci 00:25:34.469 0000:00:04.0 (8086 2021): ioatdma -> vfio-pci 00:25:34.469 0000:80:04.7 (8086 2021): ioatdma -> vfio-pci 00:25:34.469 0000:80:04.6 (8086 2021): ioatdma -> vfio-pci 00:25:34.469 0000:80:04.5 (8086 2021): ioatdma -> vfio-pci 00:25:34.469 0000:80:04.4 (8086 2021): ioatdma -> vfio-pci 00:25:34.469 0000:80:04.3 (8086 2021): ioatdma -> vfio-pci 00:25:34.469 0000:80:04.2 (8086 2021): ioatdma -> vfio-pci 00:25:34.469 0000:80:04.1 (8086 2021): ioatdma -> vfio-pci 00:25:34.469 0000:80:04.0 (8086 2021): ioatdma -> vfio-pci 00:25:35.406 0000:5e:00.0 (8086 0a54): nvme -> vfio-pci 00:25:35.406 18:11:29 nvmf_tcp.nvmf_auth_host -- host/auth.sh@28 -- # rm -f /tmp/spdk.key-null.r5r /tmp/spdk.key-null.QKj /tmp/spdk.key-sha256.JmW /tmp/spdk.key-sha384.ygI /tmp/spdk.key-sha512.vIO /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/nvme-auth.log 00:25:35.406 18:11:29 nvmf_tcp.nvmf_auth_host -- host/auth.sh@31 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:25:37.941 0000:00:04.7 (8086 2021): Already using the vfio-pci driver 00:25:37.941 0000:5e:00.0 (8086 0a54): Already using the vfio-pci driver 00:25:37.941 0000:00:04.6 (8086 2021): Already using the vfio-pci driver 00:25:37.941 0000:00:04.5 (8086 2021): Already using the vfio-pci driver 00:25:37.941 0000:00:04.4 (8086 2021): Already using the vfio-pci driver 00:25:37.941 0000:00:04.3 (8086 2021): Already using the vfio-pci driver 00:25:37.942 0000:00:04.2 (8086 2021): Already using the vfio-pci driver 00:25:37.942 0000:00:04.1 (8086 2021): Already using the vfio-pci driver 00:25:37.942 0000:00:04.0 (8086 2021): Already using the vfio-pci driver 00:25:37.942 0000:80:04.7 (8086 2021): Already using the vfio-pci driver 00:25:37.942 0000:80:04.6 (8086 2021): Already using the vfio-pci driver 00:25:37.942 0000:80:04.5 (8086 2021): Already using the vfio-pci driver 00:25:37.942 0000:80:04.4 (8086 2021): Already using the vfio-pci driver 00:25:37.942 0000:80:04.3 (8086 2021): Already using the vfio-pci driver 00:25:37.942 0000:80:04.2 (8086 2021): Already using the vfio-pci driver 00:25:37.942 0000:80:04.1 (8086 2021): Already using the vfio-pci driver 00:25:37.942 0000:80:04.0 (8086 2021): Already using the vfio-pci driver 00:25:37.942 00:25:37.942 real 0m47.703s 00:25:37.942 user 0m42.471s 00:25:37.942 sys 0m10.961s 00:25:37.942 18:11:31 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@1124 -- # xtrace_disable 00:25:37.942 18:11:31 nvmf_tcp.nvmf_auth_host -- common/autotest_common.sh@10 -- # set +x 00:25:37.942 ************************************ 00:25:37.942 END TEST nvmf_auth_host 00:25:37.942 ************************************ 00:25:37.942 18:11:31 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:25:37.942 18:11:31 nvmf_tcp -- nvmf/nvmf.sh@107 -- # [[ tcp == \t\c\p ]] 00:25:37.942 18:11:31 nvmf_tcp -- nvmf/nvmf.sh@108 -- # run_test nvmf_digest /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/digest.sh --transport=tcp 00:25:37.942 18:11:31 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:25:37.942 18:11:31 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:25:37.942 18:11:31 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:25:38.201 ************************************ 00:25:38.201 START TEST nvmf_digest 00:25:38.201 ************************************ 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/digest.sh --transport=tcp 00:25:38.201 * Looking for test storage... 00:25:38.201 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- host/digest.sh@12 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@7 -- # uname -s 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:38.201 18:11:31 nvmf_tcp.nvmf_digest -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- paths/export.sh@5 -- # export PATH 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@47 -- # : 0 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@51 -- # have_pci_nics=0 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- host/digest.sh@14 -- # nqn=nqn.2016-06.io.spdk:cnode1 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- host/digest.sh@15 -- # bperfsock=/var/tmp/bperf.sock 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- host/digest.sh@16 -- # runtime=2 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- host/digest.sh@136 -- # [[ tcp != \t\c\p ]] 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- host/digest.sh@138 -- # nvmftestinit 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@448 -- # prepare_net_devs 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@410 -- # local -g is_hw=no 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@412 -- # remove_spdk_ns 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- nvmf/common.sh@285 -- # xtrace_disable 00:25:38.202 18:11:31 nvmf_tcp.nvmf_digest -- common/autotest_common.sh@10 -- # set +x 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@291 -- # pci_devs=() 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@291 -- # local -a pci_devs 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@292 -- # pci_net_devs=() 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@293 -- # pci_drivers=() 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@293 -- # local -A pci_drivers 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@295 -- # net_devs=() 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@295 -- # local -ga net_devs 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@296 -- # e810=() 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@296 -- # local -ga e810 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@297 -- # x722=() 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@297 -- # local -ga x722 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@298 -- # mlx=() 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@298 -- # local -ga mlx 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:25:43.476 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:25:43.477 Found 0000:86:00.0 (0x8086 - 0x159b) 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:25:43.477 Found 0000:86:00.1 (0x8086 - 0x159b) 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@390 -- # [[ up == up ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:25:43.477 Found net devices under 0000:86:00.0: cvl_0_0 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@390 -- # [[ up == up ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:25:43.477 Found net devices under 0000:86:00.1: cvl_0_1 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@414 -- # is_hw=yes 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:25:43.477 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:25:43.736 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:25:43.736 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.148 ms 00:25:43.736 00:25:43.736 --- 10.0.0.2 ping statistics --- 00:25:43.736 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:43.736 rtt min/avg/max/mdev = 0.148/0.148/0.148/0.000 ms 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:25:43.736 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:25:43.736 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.205 ms 00:25:43.736 00:25:43.736 --- 10.0.0.1 ping statistics --- 00:25:43.736 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:25:43.736 rtt min/avg/max/mdev = 0.205/0.205/0.205/0.000 ms 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@422 -- # return 0 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- host/digest.sh@140 -- # trap cleanup SIGINT SIGTERM EXIT 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- host/digest.sh@141 -- # [[ 0 -eq 1 ]] 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- host/digest.sh@145 -- # run_test nvmf_digest_clean run_digest 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- common/autotest_common.sh@1105 -- # xtrace_disable 00:25:43.736 18:11:37 nvmf_tcp.nvmf_digest -- common/autotest_common.sh@10 -- # set +x 00:25:43.736 ************************************ 00:25:43.736 START TEST nvmf_digest_clean 00:25:43.736 ************************************ 00:25:43.737 18:11:37 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@1123 -- # run_digest 00:25:43.737 18:11:37 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@120 -- # local dsa_initiator 00:25:43.737 18:11:37 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@121 -- # [[ '' == \d\s\a\_\i\n\i\t\i\a\t\o\r ]] 00:25:43.737 18:11:37 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@121 -- # dsa_initiator=false 00:25:43.737 18:11:37 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@123 -- # tgt_params=("--wait-for-rpc") 00:25:43.737 18:11:37 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@124 -- # nvmfappstart --wait-for-rpc 00:25:43.737 18:11:37 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:25:43.737 18:11:37 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@722 -- # xtrace_disable 00:25:43.737 18:11:37 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:25:43.737 18:11:37 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@481 -- # nvmfpid=730029 00:25:43.737 18:11:37 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@482 -- # waitforlisten 730029 00:25:43.737 18:11:37 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc 00:25:43.737 18:11:37 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@829 -- # '[' -z 730029 ']' 00:25:43.737 18:11:37 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:25:43.737 18:11:37 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@834 -- # local max_retries=100 00:25:43.737 18:11:37 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:25:43.737 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:25:43.737 18:11:37 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@838 -- # xtrace_disable 00:25:43.737 18:11:37 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:25:43.996 [2024-07-15 18:11:37.479710] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:25:43.996 [2024-07-15 18:11:37.479750] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:25:43.996 EAL: No free 2048 kB hugepages reported on node 1 00:25:43.996 [2024-07-15 18:11:37.536098] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:43.996 [2024-07-15 18:11:37.614465] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:25:43.996 [2024-07-15 18:11:37.614504] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:25:43.996 [2024-07-15 18:11:37.614511] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:25:43.996 [2024-07-15 18:11:37.614517] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:25:43.996 [2024-07-15 18:11:37.614522] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:25:43.996 [2024-07-15 18:11:37.614556] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:25:44.565 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:25:44.565 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@862 -- # return 0 00:25:44.565 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:25:44.565 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@728 -- # xtrace_disable 00:25:44.565 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@125 -- # [[ '' == \d\s\a\_\t\a\r\g\e\t ]] 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@126 -- # common_target_config 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@43 -- # rpc_cmd 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@559 -- # xtrace_disable 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:25:44.825 null0 00:25:44.825 [2024-07-15 18:11:38.402731] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:25:44.825 [2024-07-15 18:11:38.426913] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@128 -- # run_bperf randread 4096 128 false 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # rw=randread 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # bs=4096 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # qd=128 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # scan_dsa=false 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@83 -- # bperfpid=730262 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@84 -- # waitforlisten 730262 /var/tmp/bperf.sock 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@829 -- # '[' -z 730262 ']' 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bperf.sock 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@834 -- # local max_retries=100 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:25:44.825 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@838 -- # xtrace_disable 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:25:44.825 18:11:38 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 4096 -t 2 -q 128 -z --wait-for-rpc 00:25:44.825 [2024-07-15 18:11:38.477104] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:25:44.825 [2024-07-15 18:11:38.477145] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid730262 ] 00:25:44.825 EAL: No free 2048 kB hugepages reported on node 1 00:25:44.825 [2024-07-15 18:11:38.529931] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:45.084 [2024-07-15 18:11:38.609704] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:25:45.653 18:11:39 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:25:45.653 18:11:39 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@862 -- # return 0 00:25:45.653 18:11:39 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@86 -- # false 00:25:45.653 18:11:39 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:25:45.653 18:11:39 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:25:45.912 18:11:39 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:25:45.912 18:11:39 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:25:46.172 nvme0n1 00:25:46.172 18:11:39 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@92 -- # bperf_py perform_tests 00:25:46.172 18:11:39 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:25:46.431 Running I/O for 2 seconds... 00:25:48.351 00:25:48.351 Latency(us) 00:25:48.351 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:25:48.351 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 128, IO size: 4096) 00:25:48.351 nvme0n1 : 2.00 26379.26 103.04 0.00 0.00 4847.43 2194.03 13848.04 00:25:48.351 =================================================================================================================== 00:25:48.351 Total : 26379.26 103.04 0.00 0.00 4847.43 2194.03 13848.04 00:25:48.351 0 00:25:48.351 18:11:41 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:25:48.351 18:11:41 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # get_accel_stats 00:25:48.351 18:11:41 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:25:48.351 18:11:41 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@37 -- # jq -rc '.operations[] 00:25:48.351 | select(.opcode=="crc32c") 00:25:48.351 | "\(.module_name) \(.executed)"' 00:25:48.351 18:11:41 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:25:48.610 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # false 00:25:48.610 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # exp_module=software 00:25:48.610 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:25:48.610 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:25:48.610 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@98 -- # killprocess 730262 00:25:48.610 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@948 -- # '[' -z 730262 ']' 00:25:48.610 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@952 -- # kill -0 730262 00:25:48.610 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@953 -- # uname 00:25:48.610 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:25:48.610 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 730262 00:25:48.610 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:25:48.610 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:25:48.610 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@966 -- # echo 'killing process with pid 730262' 00:25:48.610 killing process with pid 730262 00:25:48.610 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@967 -- # kill 730262 00:25:48.610 Received shutdown signal, test time was about 2.000000 seconds 00:25:48.610 00:25:48.610 Latency(us) 00:25:48.610 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:25:48.610 =================================================================================================================== 00:25:48.610 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:25:48.610 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@972 -- # wait 730262 00:25:48.868 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@129 -- # run_bperf randread 131072 16 false 00:25:48.869 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:25:48.869 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:25:48.869 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # rw=randread 00:25:48.869 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # bs=131072 00:25:48.869 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # qd=16 00:25:48.869 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # scan_dsa=false 00:25:48.869 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@83 -- # bperfpid=730956 00:25:48.869 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@84 -- # waitforlisten 730956 /var/tmp/bperf.sock 00:25:48.869 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 131072 -t 2 -q 16 -z --wait-for-rpc 00:25:48.869 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@829 -- # '[' -z 730956 ']' 00:25:48.869 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bperf.sock 00:25:48.869 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@834 -- # local max_retries=100 00:25:48.869 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:25:48.869 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:25:48.869 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@838 -- # xtrace_disable 00:25:48.869 18:11:42 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:25:48.869 [2024-07-15 18:11:42.393871] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:25:48.869 [2024-07-15 18:11:42.393919] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid730956 ] 00:25:48.869 I/O size of 131072 is greater than zero copy threshold (65536). 00:25:48.869 Zero copy mechanism will not be used. 00:25:48.869 EAL: No free 2048 kB hugepages reported on node 1 00:25:48.869 [2024-07-15 18:11:42.447553] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:48.869 [2024-07-15 18:11:42.515301] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:25:49.507 18:11:43 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:25:49.507 18:11:43 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@862 -- # return 0 00:25:49.507 18:11:43 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@86 -- # false 00:25:49.507 18:11:43 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:25:49.507 18:11:43 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:25:49.765 18:11:43 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:25:49.765 18:11:43 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:25:50.023 nvme0n1 00:25:50.023 18:11:43 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@92 -- # bperf_py perform_tests 00:25:50.023 18:11:43 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:25:50.282 I/O size of 131072 is greater than zero copy threshold (65536). 00:25:50.282 Zero copy mechanism will not be used. 00:25:50.282 Running I/O for 2 seconds... 00:25:52.188 00:25:52.188 Latency(us) 00:25:52.188 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:25:52.188 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 16, IO size: 131072) 00:25:52.188 nvme0n1 : 2.00 4676.66 584.58 0.00 0.00 3418.85 651.80 5727.28 00:25:52.188 =================================================================================================================== 00:25:52.188 Total : 4676.66 584.58 0.00 0.00 3418.85 651.80 5727.28 00:25:52.188 0 00:25:52.188 18:11:45 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:25:52.188 18:11:45 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # get_accel_stats 00:25:52.188 18:11:45 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:25:52.188 18:11:45 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@37 -- # jq -rc '.operations[] 00:25:52.188 | select(.opcode=="crc32c") 00:25:52.188 | "\(.module_name) \(.executed)"' 00:25:52.188 18:11:45 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:25:52.447 18:11:45 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # false 00:25:52.447 18:11:45 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # exp_module=software 00:25:52.447 18:11:45 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:25:52.447 18:11:45 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:25:52.447 18:11:45 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@98 -- # killprocess 730956 00:25:52.447 18:11:45 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@948 -- # '[' -z 730956 ']' 00:25:52.447 18:11:45 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@952 -- # kill -0 730956 00:25:52.447 18:11:45 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@953 -- # uname 00:25:52.447 18:11:45 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:25:52.447 18:11:45 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 730956 00:25:52.447 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:25:52.447 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:25:52.447 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@966 -- # echo 'killing process with pid 730956' 00:25:52.447 killing process with pid 730956 00:25:52.447 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@967 -- # kill 730956 00:25:52.447 Received shutdown signal, test time was about 2.000000 seconds 00:25:52.447 00:25:52.447 Latency(us) 00:25:52.447 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:25:52.447 =================================================================================================================== 00:25:52.447 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:25:52.447 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@972 -- # wait 730956 00:25:52.706 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@130 -- # run_bperf randwrite 4096 128 false 00:25:52.706 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:25:52.706 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:25:52.706 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # rw=randwrite 00:25:52.706 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # bs=4096 00:25:52.706 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # qd=128 00:25:52.706 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # scan_dsa=false 00:25:52.706 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@83 -- # bperfpid=731602 00:25:52.706 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@84 -- # waitforlisten 731602 /var/tmp/bperf.sock 00:25:52.706 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 4096 -t 2 -q 128 -z --wait-for-rpc 00:25:52.706 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@829 -- # '[' -z 731602 ']' 00:25:52.706 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bperf.sock 00:25:52.706 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@834 -- # local max_retries=100 00:25:52.706 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:25:52.706 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:25:52.706 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@838 -- # xtrace_disable 00:25:52.706 18:11:46 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:25:52.706 [2024-07-15 18:11:46.262738] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:25:52.706 [2024-07-15 18:11:46.262786] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid731602 ] 00:25:52.706 EAL: No free 2048 kB hugepages reported on node 1 00:25:52.706 [2024-07-15 18:11:46.317170] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:52.706 [2024-07-15 18:11:46.396433] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:25:53.644 18:11:47 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:25:53.644 18:11:47 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@862 -- # return 0 00:25:53.644 18:11:47 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@86 -- # false 00:25:53.644 18:11:47 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:25:53.644 18:11:47 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:25:53.644 18:11:47 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:25:53.644 18:11:47 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:25:54.211 nvme0n1 00:25:54.211 18:11:47 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@92 -- # bperf_py perform_tests 00:25:54.211 18:11:47 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:25:54.212 Running I/O for 2 seconds... 00:25:56.116 00:25:56.116 Latency(us) 00:25:56.116 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:25:56.116 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:25:56.116 nvme0n1 : 2.00 27458.22 107.26 0.00 0.00 4653.89 2208.28 14702.86 00:25:56.116 =================================================================================================================== 00:25:56.116 Total : 27458.22 107.26 0.00 0.00 4653.89 2208.28 14702.86 00:25:56.116 0 00:25:56.116 18:11:49 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:25:56.116 18:11:49 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # get_accel_stats 00:25:56.116 18:11:49 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:25:56.116 18:11:49 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@37 -- # jq -rc '.operations[] 00:25:56.116 | select(.opcode=="crc32c") 00:25:56.116 | "\(.module_name) \(.executed)"' 00:25:56.116 18:11:49 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:25:56.377 18:11:49 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # false 00:25:56.377 18:11:49 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # exp_module=software 00:25:56.377 18:11:49 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:25:56.377 18:11:49 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:25:56.377 18:11:49 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@98 -- # killprocess 731602 00:25:56.377 18:11:49 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@948 -- # '[' -z 731602 ']' 00:25:56.377 18:11:49 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@952 -- # kill -0 731602 00:25:56.377 18:11:49 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@953 -- # uname 00:25:56.377 18:11:49 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:25:56.377 18:11:49 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 731602 00:25:56.377 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:25:56.377 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:25:56.377 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@966 -- # echo 'killing process with pid 731602' 00:25:56.377 killing process with pid 731602 00:25:56.377 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@967 -- # kill 731602 00:25:56.377 Received shutdown signal, test time was about 2.000000 seconds 00:25:56.377 00:25:56.377 Latency(us) 00:25:56.377 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:25:56.377 =================================================================================================================== 00:25:56.377 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:25:56.377 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@972 -- # wait 731602 00:25:56.636 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@131 -- # run_bperf randwrite 131072 16 false 00:25:56.636 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@77 -- # local rw bs qd scan_dsa 00:25:56.636 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@78 -- # local acc_module acc_executed exp_module 00:25:56.636 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # rw=randwrite 00:25:56.636 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # bs=131072 00:25:56.636 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # qd=16 00:25:56.636 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@80 -- # scan_dsa=false 00:25:56.636 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@83 -- # bperfpid=732154 00:25:56.636 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@84 -- # waitforlisten 732154 /var/tmp/bperf.sock 00:25:56.636 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@82 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 131072 -t 2 -q 16 -z --wait-for-rpc 00:25:56.636 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@829 -- # '[' -z 732154 ']' 00:25:56.636 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bperf.sock 00:25:56.636 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@834 -- # local max_retries=100 00:25:56.636 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:25:56.636 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:25:56.636 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@838 -- # xtrace_disable 00:25:56.636 18:11:50 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:25:56.636 [2024-07-15 18:11:50.260223] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:25:56.636 [2024-07-15 18:11:50.260278] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid732154 ] 00:25:56.636 I/O size of 131072 is greater than zero copy threshold (65536). 00:25:56.636 Zero copy mechanism will not be used. 00:25:56.636 EAL: No free 2048 kB hugepages reported on node 1 00:25:56.636 [2024-07-15 18:11:50.315457] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:25:56.894 [2024-07-15 18:11:50.389789] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:25:57.461 18:11:51 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:25:57.461 18:11:51 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@862 -- # return 0 00:25:57.461 18:11:51 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@86 -- # false 00:25:57.461 18:11:51 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@87 -- # bperf_rpc framework_start_init 00:25:57.461 18:11:51 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:25:57.720 18:11:51 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@89 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:25:57.720 18:11:51 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:25:57.979 nvme0n1 00:25:57.979 18:11:51 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@92 -- # bperf_py perform_tests 00:25:57.979 18:11:51 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:25:57.979 I/O size of 131072 is greater than zero copy threshold (65536). 00:25:57.979 Zero copy mechanism will not be used. 00:25:57.979 Running I/O for 2 seconds... 00:25:59.885 00:25:59.885 Latency(us) 00:25:59.885 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:25:59.885 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 16, IO size: 131072) 00:25:59.885 nvme0n1 : 2.00 5560.11 695.01 0.00 0.00 2872.48 1809.36 8206.25 00:25:59.885 =================================================================================================================== 00:25:59.885 Total : 5560.11 695.01 0.00 0.00 2872.48 1809.36 8206.25 00:25:59.885 0 00:26:00.144 18:11:53 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # read -r acc_module acc_executed 00:26:00.144 18:11:53 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@93 -- # get_accel_stats 00:26:00.144 18:11:53 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@36 -- # bperf_rpc accel_get_stats 00:26:00.144 18:11:53 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@37 -- # jq -rc '.operations[] 00:26:00.144 | select(.opcode=="crc32c") 00:26:00.144 | "\(.module_name) \(.executed)"' 00:26:00.144 18:11:53 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock accel_get_stats 00:26:00.144 18:11:53 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # false 00:26:00.144 18:11:53 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@94 -- # exp_module=software 00:26:00.144 18:11:53 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@95 -- # (( acc_executed > 0 )) 00:26:00.144 18:11:53 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@96 -- # [[ software == \s\o\f\t\w\a\r\e ]] 00:26:00.144 18:11:53 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@98 -- # killprocess 732154 00:26:00.144 18:11:53 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@948 -- # '[' -z 732154 ']' 00:26:00.144 18:11:53 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@952 -- # kill -0 732154 00:26:00.144 18:11:53 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@953 -- # uname 00:26:00.144 18:11:53 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:26:00.144 18:11:53 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 732154 00:26:00.144 18:11:53 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:26:00.144 18:11:53 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:26:00.144 18:11:53 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@966 -- # echo 'killing process with pid 732154' 00:26:00.144 killing process with pid 732154 00:26:00.144 18:11:53 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@967 -- # kill 732154 00:26:00.144 Received shutdown signal, test time was about 2.000000 seconds 00:26:00.144 00:26:00.144 Latency(us) 00:26:00.144 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:00.144 =================================================================================================================== 00:26:00.144 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:26:00.144 18:11:53 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@972 -- # wait 732154 00:26:00.404 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- host/digest.sh@132 -- # killprocess 730029 00:26:00.404 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@948 -- # '[' -z 730029 ']' 00:26:00.404 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@952 -- # kill -0 730029 00:26:00.404 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@953 -- # uname 00:26:00.404 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:26:00.404 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 730029 00:26:00.404 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:26:00.404 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:26:00.404 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@966 -- # echo 'killing process with pid 730029' 00:26:00.404 killing process with pid 730029 00:26:00.404 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@967 -- # kill 730029 00:26:00.404 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@972 -- # wait 730029 00:26:00.664 00:26:00.664 real 0m16.820s 00:26:00.664 user 0m32.324s 00:26:00.664 sys 0m4.339s 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@1124 -- # xtrace_disable 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_clean -- common/autotest_common.sh@10 -- # set +x 00:26:00.664 ************************************ 00:26:00.664 END TEST nvmf_digest_clean 00:26:00.664 ************************************ 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest -- common/autotest_common.sh@1142 -- # return 0 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest -- host/digest.sh@147 -- # run_test nvmf_digest_error run_digest_error 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest -- common/autotest_common.sh@1105 -- # xtrace_disable 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest -- common/autotest_common.sh@10 -- # set +x 00:26:00.664 ************************************ 00:26:00.664 START TEST nvmf_digest_error 00:26:00.664 ************************************ 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@1123 -- # run_digest_error 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@102 -- # nvmfappstart --wait-for-rpc 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@722 -- # xtrace_disable 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@481 -- # nvmfpid=732859 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@482 -- # waitforlisten 732859 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF --wait-for-rpc 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@829 -- # '[' -z 732859 ']' 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@834 -- # local max_retries=100 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:00.664 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@838 -- # xtrace_disable 00:26:00.664 18:11:54 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:26:00.664 [2024-07-15 18:11:54.369743] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:26:00.664 [2024-07-15 18:11:54.369786] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:00.924 EAL: No free 2048 kB hugepages reported on node 1 00:26:00.924 [2024-07-15 18:11:54.428749] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:00.924 [2024-07-15 18:11:54.499108] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:26:00.924 [2024-07-15 18:11:54.499146] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:26:00.924 [2024-07-15 18:11:54.499153] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:26:00.924 [2024-07-15 18:11:54.499159] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:26:00.924 [2024-07-15 18:11:54.499165] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:26:00.924 [2024-07-15 18:11:54.499182] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:26:01.491 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:26:01.491 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@862 -- # return 0 00:26:01.491 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:26:01.491 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@728 -- # xtrace_disable 00:26:01.491 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:26:01.491 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:26:01.491 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@104 -- # rpc_cmd accel_assign_opc -o crc32c -m error 00:26:01.491 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:01.491 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:26:01.491 [2024-07-15 18:11:55.217273] accel_rpc.c: 167:rpc_accel_assign_opc: *NOTICE*: Operation crc32c will be assigned to module error 00:26:01.751 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:01.751 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@105 -- # common_target_config 00:26:01.751 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@43 -- # rpc_cmd 00:26:01.751 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:01.751 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:26:01.751 null0 00:26:01.751 [2024-07-15 18:11:55.305283] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:26:01.751 [2024-07-15 18:11:55.329471] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:01.751 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:01.751 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@108 -- # run_bperf_err randread 4096 128 00:26:01.751 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@54 -- # local rw bs qd 00:26:01.751 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # rw=randread 00:26:01.751 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # bs=4096 00:26:01.751 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # qd=128 00:26:01.751 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@58 -- # bperfpid=733104 00:26:01.751 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@60 -- # waitforlisten 733104 /var/tmp/bperf.sock 00:26:01.751 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 4096 -t 2 -q 128 -z 00:26:01.751 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@829 -- # '[' -z 733104 ']' 00:26:01.751 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bperf.sock 00:26:01.751 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@834 -- # local max_retries=100 00:26:01.751 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:26:01.751 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:26:01.751 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@838 -- # xtrace_disable 00:26:01.751 18:11:55 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:26:01.751 [2024-07-15 18:11:55.381255] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:26:01.751 [2024-07-15 18:11:55.381295] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid733104 ] 00:26:01.751 EAL: No free 2048 kB hugepages reported on node 1 00:26:01.751 [2024-07-15 18:11:55.434790] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:02.010 [2024-07-15 18:11:55.514001] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:26:02.582 18:11:56 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:26:02.582 18:11:56 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@862 -- # return 0 00:26:02.582 18:11:56 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:26:02.582 18:11:56 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:26:02.841 18:11:56 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:26:02.841 18:11:56 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:02.841 18:11:56 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:26:02.841 18:11:56 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:02.841 18:11:56 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:26:02.841 18:11:56 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:26:03.102 nvme0n1 00:26:03.102 18:11:56 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 256 00:26:03.102 18:11:56 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:03.102 18:11:56 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:26:03.102 18:11:56 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:03.102 18:11:56 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@69 -- # bperf_py perform_tests 00:26:03.102 18:11:56 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:26:03.102 Running I/O for 2 seconds... 00:26:03.102 [2024-07-15 18:11:56.716631] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.102 [2024-07-15 18:11:56.716664] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:15845 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.102 [2024-07-15 18:11:56.716675] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.102 [2024-07-15 18:11:56.725347] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.102 [2024-07-15 18:11:56.725370] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:23564 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.102 [2024-07-15 18:11:56.725379] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:32 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.102 [2024-07-15 18:11:56.735542] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.102 [2024-07-15 18:11:56.735564] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:9036 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.102 [2024-07-15 18:11:56.735573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:28 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.102 [2024-07-15 18:11:56.745538] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.102 [2024-07-15 18:11:56.745558] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:2538 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.102 [2024-07-15 18:11:56.745567] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:63 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.102 [2024-07-15 18:11:56.754173] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.102 [2024-07-15 18:11:56.754194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:21743 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.102 [2024-07-15 18:11:56.754203] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:61 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.102 [2024-07-15 18:11:56.766738] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.102 [2024-07-15 18:11:56.766759] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:82 nsid:1 lba:8448 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.102 [2024-07-15 18:11:56.766767] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:82 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.102 [2024-07-15 18:11:56.777627] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.102 [2024-07-15 18:11:56.777647] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:3788 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.102 [2024-07-15 18:11:56.777656] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:75 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.102 [2024-07-15 18:11:56.788069] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.102 [2024-07-15 18:11:56.788088] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:13060 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.102 [2024-07-15 18:11:56.788096] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:89 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.102 [2024-07-15 18:11:56.795949] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.102 [2024-07-15 18:11:56.795968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:11426 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.102 [2024-07-15 18:11:56.795976] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:23 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.102 [2024-07-15 18:11:56.808649] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.102 [2024-07-15 18:11:56.808669] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:17262 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.102 [2024-07-15 18:11:56.808677] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.102 [2024-07-15 18:11:56.818931] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.102 [2024-07-15 18:11:56.818951] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:4009 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.102 [2024-07-15 18:11:56.818962] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:58 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.102 [2024-07-15 18:11:56.826948] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.102 [2024-07-15 18:11:56.826967] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:15011 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.102 [2024-07-15 18:11:56.826975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:58 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.361 [2024-07-15 18:11:56.838837] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.361 [2024-07-15 18:11:56.838859] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:9025 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.361 [2024-07-15 18:11:56.838868] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:107 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.361 [2024-07-15 18:11:56.850403] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.361 [2024-07-15 18:11:56.850423] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:6185 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.361 [2024-07-15 18:11:56.850431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:85 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.361 [2024-07-15 18:11:56.858943] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.361 [2024-07-15 18:11:56.858963] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:6303 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:56.858971] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:22 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:56.870627] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:56.870647] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:5074 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:56.870655] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:56.882408] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:56.882428] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:21890 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:56.882436] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:34 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:56.890857] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:56.890876] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:19953 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:56.890884] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:77 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:56.901997] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:56.902017] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:2525 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:56.902025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:51 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:56.913303] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:56.913326] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:19079 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:56.913335] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:84 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:56.922902] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:56.922922] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:24715 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:56.922930] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:98 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:56.931235] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:56.931254] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:25463 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:56.931262] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:47 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:56.944451] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:56.944470] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:18176 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:56.944478] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:48 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:56.953853] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:56.953873] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:11696 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:56.953881] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:67 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:56.965064] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:56.965084] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:21490 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:56.965092] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:56.972996] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:56.973015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:1374 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:56.973023] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:30 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:56.983443] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:56.983463] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:21780 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:56.983471] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:27 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:56.995182] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:56.995202] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:4386 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:56.995211] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:62 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:57.006523] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:57.006543] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:19389 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:57.006551] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:57.016464] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:57.016484] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:6199 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:57.016491] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:115 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:57.025189] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:57.025208] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:14489 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:57.025216] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:90 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:57.034392] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:57.034412] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:16298 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:57.034420] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:88 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:57.045699] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:57.045720] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:6750 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:57.045728] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:34 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:57.054222] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:57.054248] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:8835 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:57.054256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:116 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:57.064383] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:57.064402] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:15227 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:57.064410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:27 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:57.074835] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:57.074855] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:23807 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:57.074862] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:89 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.362 [2024-07-15 18:11:57.083251] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.362 [2024-07-15 18:11:57.083270] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:17087 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.362 [2024-07-15 18:11:57.083282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.093652] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.093672] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:21172 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.093680] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:89 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.102690] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.102709] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:8983 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.102716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.111580] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.111599] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:24524 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.111607] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.121691] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.121710] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:23090 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.121718] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.131322] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.131342] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:11564 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.131350] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:76 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.140167] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.140186] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:21051 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.140195] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:31 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.152111] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.152131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:12202 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.152138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:92 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.162877] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.162897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:5760 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.162904] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:18 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.176031] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.176052] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:8763 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.176059] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:23 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.184272] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.184291] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:19238 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.184299] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:21 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.194094] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.194114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:19235 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.194122] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:55 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.204270] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.204290] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:19419 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.204298] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:99 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.212439] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.212459] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:5156 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.212466] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:79 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.222327] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.222347] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:16772 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.222354] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:76 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.232068] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.232088] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:23984 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.232096] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:96 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.241744] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.241764] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:10987 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.241772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.250234] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.250254] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:19735 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.250265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:90 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.261277] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.261296] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:9330 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.261304] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:58 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.269102] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.269123] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:16827 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.269131] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:84 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.280488] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.280508] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:7875 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.280516] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:35 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.621 [2024-07-15 18:11:57.288658] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.621 [2024-07-15 18:11:57.288678] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:20791 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.621 [2024-07-15 18:11:57.288686] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.622 [2024-07-15 18:11:57.298050] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.622 [2024-07-15 18:11:57.298071] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:9343 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.622 [2024-07-15 18:11:57.298078] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.622 [2024-07-15 18:11:57.309254] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.622 [2024-07-15 18:11:57.309275] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:10908 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.622 [2024-07-15 18:11:57.309283] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:84 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.622 [2024-07-15 18:11:57.317589] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.622 [2024-07-15 18:11:57.317609] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:81 nsid:1 lba:23275 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.622 [2024-07-15 18:11:57.317617] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:81 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.622 [2024-07-15 18:11:57.328984] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.622 [2024-07-15 18:11:57.329004] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:16533 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.622 [2024-07-15 18:11:57.329013] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.622 [2024-07-15 18:11:57.340010] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.622 [2024-07-15 18:11:57.340033] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:14759 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.622 [2024-07-15 18:11:57.340041] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:30 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.349205] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.349230] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:1694 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.349239] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:84 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.358247] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.358266] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:1367 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.358274] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:41 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.367552] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.367572] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:7474 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.367580] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:106 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.376548] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.376568] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:24921 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.376575] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:97 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.387393] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.387413] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:20731 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.387420] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:50 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.397221] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.397248] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:14445 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.397255] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:77 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.404989] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.405009] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:66 nsid:1 lba:24832 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.405017] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:66 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.415260] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.415280] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:24689 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.415288] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.424774] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.424794] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:25562 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.424802] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:71 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.433260] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.433280] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:17667 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.433287] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:37 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.442775] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.442798] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:9184 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.442806] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.452324] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.452343] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:15677 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.452351] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.462060] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.462080] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:21080 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.462088] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:44 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.470940] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.470959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:7261 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.470967] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:60 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.479588] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.479608] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:11684 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.479615] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:47 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.489704] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.489723] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:15739 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.489732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:83 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.499871] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.499891] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:12923 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.499903] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:64 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.508329] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.508351] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:4659 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.508359] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:28 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.519161] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.519180] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:1038 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.519188] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.529995] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.530014] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:89 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.530022] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:21 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.538846] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.538866] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:17216 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.538874] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:17 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.550041] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.550061] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:680 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.550069] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:63 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.557861] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.557881] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:13518 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.557888] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.568105] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.568124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:7283 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.568132] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:60 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.578188] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.578208] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:1677 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.578216] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:70 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.586388] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.586410] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:24659 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.586418] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:113 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.595471] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.595490] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:1928 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.595498] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:125 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:03.881 [2024-07-15 18:11:57.606216] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:03.881 [2024-07-15 18:11:57.606242] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:10712 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:03.881 [2024-07-15 18:11:57.606251] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:92 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.139 [2024-07-15 18:11:57.615108] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.139 [2024-07-15 18:11:57.615128] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:14141 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.139 [2024-07-15 18:11:57.615136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:58 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.139 [2024-07-15 18:11:57.624389] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.139 [2024-07-15 18:11:57.624408] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:10809 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.139 [2024-07-15 18:11:57.624417] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:90 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.634194] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.634215] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:11210 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.634223] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.644215] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.644241] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:4100 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.644249] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:53 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.652873] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.652893] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:22298 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.652901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.662281] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.662301] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:13775 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.662309] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:119 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.672560] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.672580] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:13080 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.672588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.680940] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.680960] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:24549 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.680968] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:113 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.690635] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.690655] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:2269 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.690662] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.700360] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.700379] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:15611 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.700387] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:62 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.709721] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.709740] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:21219 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.709748] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.719479] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.719498] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:5593 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.719506] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.728278] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.728298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:9776 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.728305] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:29 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.740123] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.740144] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:21522 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.740152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:44 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.748864] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.748888] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:18060 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.748896] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.758449] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.758469] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:23469 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.758477] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:100 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.768838] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.768857] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:25004 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.768865] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:30 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.777609] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.777629] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:5083 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.777637] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:91 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.786570] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.786589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:15717 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.786597] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:85 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.796117] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.796137] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:6816 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.796145] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:30 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.805744] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.805763] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:13531 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.805772] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.814461] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.814481] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:8074 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.814489] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.824211] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.824235] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:15536 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.824243] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:95 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.833410] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.833430] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:8541 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.833438] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:105 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.843293] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.843312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:14902 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.843320] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.853342] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.853362] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:24261 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.853369] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:27 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.140 [2024-07-15 18:11:57.861231] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.140 [2024-07-15 18:11:57.861250] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:10653 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.140 [2024-07-15 18:11:57.861258] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:20 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.398 [2024-07-15 18:11:57.872046] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.398 [2024-07-15 18:11:57.872066] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:8707 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.398 [2024-07-15 18:11:57.872074] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:21 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.398 [2024-07-15 18:11:57.882194] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.398 [2024-07-15 18:11:57.882213] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:10021 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.398 [2024-07-15 18:11:57.882221] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:30 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.398 [2024-07-15 18:11:57.890950] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.398 [2024-07-15 18:11:57.890969] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:13217 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.398 [2024-07-15 18:11:57.890977] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:33 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.398 [2024-07-15 18:11:57.900592] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.398 [2024-07-15 18:11:57.900612] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:8014 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.398 [2024-07-15 18:11:57.900620] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:99 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.398 [2024-07-15 18:11:57.910480] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.398 [2024-07-15 18:11:57.910500] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:8223 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.398 [2024-07-15 18:11:57.910511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:54 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.398 [2024-07-15 18:11:57.921601] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.398 [2024-07-15 18:11:57.921620] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:16901 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.398 [2024-07-15 18:11:57.921628] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:108 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.398 [2024-07-15 18:11:57.930768] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.398 [2024-07-15 18:11:57.930789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:165 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.398 [2024-07-15 18:11:57.930796] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:91 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.398 [2024-07-15 18:11:57.941545] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.398 [2024-07-15 18:11:57.941565] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:5334 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.398 [2024-07-15 18:11:57.941572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:32 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.398 [2024-07-15 18:11:57.949934] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.398 [2024-07-15 18:11:57.949953] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:1712 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.398 [2024-07-15 18:11:57.949961] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:96 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.398 [2024-07-15 18:11:57.960933] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.398 [2024-07-15 18:11:57.960952] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:3101 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.398 [2024-07-15 18:11:57.960960] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:98 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.398 [2024-07-15 18:11:57.972570] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.398 [2024-07-15 18:11:57.972590] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:3781 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.398 [2024-07-15 18:11:57.972597] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.398 [2024-07-15 18:11:57.982591] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.398 [2024-07-15 18:11:57.982610] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:15157 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.398 [2024-07-15 18:11:57.982618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:18 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.398 [2024-07-15 18:11:57.991093] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.398 [2024-07-15 18:11:57.991112] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:12833 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.398 [2024-07-15 18:11:57.991120] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:25 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.398 [2024-07-15 18:11:58.000886] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.398 [2024-07-15 18:11:58.000909] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:2983 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.398 [2024-07-15 18:11:58.000917] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:125 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.398 [2024-07-15 18:11:58.010731] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.398 [2024-07-15 18:11:58.010750] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:5991 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.398 [2024-07-15 18:11:58.010757] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:20 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.398 [2024-07-15 18:11:58.019574] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.398 [2024-07-15 18:11:58.019594] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:18269 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.398 [2024-07-15 18:11:58.019601] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:18 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.398 [2024-07-15 18:11:58.029034] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.399 [2024-07-15 18:11:58.029053] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:23121 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.399 [2024-07-15 18:11:58.029061] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:122 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.399 [2024-07-15 18:11:58.038098] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.399 [2024-07-15 18:11:58.038117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:18893 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.399 [2024-07-15 18:11:58.038125] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:91 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.399 [2024-07-15 18:11:58.047422] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.399 [2024-07-15 18:11:58.047442] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:11873 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.399 [2024-07-15 18:11:58.047450] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:62 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.399 [2024-07-15 18:11:58.057046] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.399 [2024-07-15 18:11:58.057067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:9262 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.399 [2024-07-15 18:11:58.057075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:91 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.399 [2024-07-15 18:11:58.066784] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.399 [2024-07-15 18:11:58.066804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:6285 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.399 [2024-07-15 18:11:58.066812] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:60 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.399 [2024-07-15 18:11:58.075295] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.399 [2024-07-15 18:11:58.075314] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:7336 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.399 [2024-07-15 18:11:58.075322] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:29 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.399 [2024-07-15 18:11:58.085107] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.399 [2024-07-15 18:11:58.085126] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:3063 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.399 [2024-07-15 18:11:58.085134] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:32 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.399 [2024-07-15 18:11:58.094254] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.399 [2024-07-15 18:11:58.094273] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:4220 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.399 [2024-07-15 18:11:58.094281] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:108 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.399 [2024-07-15 18:11:58.102580] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.399 [2024-07-15 18:11:58.102599] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:4499 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.399 [2024-07-15 18:11:58.102607] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:117 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.399 [2024-07-15 18:11:58.111891] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.399 [2024-07-15 18:11:58.111910] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:14657 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.399 [2024-07-15 18:11:58.111918] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:60 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.399 [2024-07-15 18:11:58.122134] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.399 [2024-07-15 18:11:58.122153] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:22240 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.399 [2024-07-15 18:11:58.122161] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:105 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.659 [2024-07-15 18:11:58.130881] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.659 [2024-07-15 18:11:58.130902] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:17672 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.130910] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:99 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.140704] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.140723] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:13201 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.140732] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.150499] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.150517] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:7571 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.150525] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:108 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.158510] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.158529] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:14913 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.158540] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:49 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.168262] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.168281] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:21122 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.168289] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:68 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.177806] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.177825] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:12323 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.177833] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:108 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.187158] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.187177] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:18054 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.187185] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:36 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.196007] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.196027] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:943 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.196034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:42 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.205472] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.205491] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:25485 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.205499] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:64 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.215265] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.215284] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:5719 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.215292] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:92 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.224053] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.224072] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:4596 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.224080] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:60 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.233615] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.233634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:6449 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.233641] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:103 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.245784] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.245804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:8889 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.245812] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.255555] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.255574] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:24747 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.255582] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:22 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.264938] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.264956] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:12455 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.264964] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:108 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.274413] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.274432] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:24379 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.274440] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:49 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.282578] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.282597] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:10377 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.282604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:121 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.292738] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.292757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:16140 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.292765] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:75 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.301165] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.301184] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:12376 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.301192] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:121 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.312253] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.312273] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:11954 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.312282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.320073] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.320092] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:14701 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.320103] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.330409] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.330428] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:24792 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.330436] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:33 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.341524] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.341544] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:102 nsid:1 lba:13744 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.341552] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:102 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.349712] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.349731] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:21230 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.349738] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:71 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.359703] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.359723] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:4040 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.359731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:115 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.369453] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.369472] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:22236 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.369480] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:36 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.660 [2024-07-15 18:11:58.378713] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.660 [2024-07-15 18:11:58.378733] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:4682 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.660 [2024-07-15 18:11:58.378741] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:90 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.388548] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.388568] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:11186 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.388577] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:93 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.398163] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.398184] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:20971 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.398192] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:53 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.406034] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.406059] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:11606 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.406067] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.416287] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.416306] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:21167 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.416314] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:51 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.425062] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.425081] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:10037 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.425089] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:32 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.434769] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.434789] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:18455 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.434796] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.444531] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.444551] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:2547 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.444558] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:22 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.454074] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.454092] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:13398 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.454100] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:104 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.462436] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.462455] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:14363 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.462463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:105 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.471993] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.472012] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:8830 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.472020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.481297] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.481316] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:3537 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.481324] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:65 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.490792] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.490812] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:24493 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.490821] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:32 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.499972] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.499992] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:1916 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.500000] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:25 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.508853] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.508872] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:834 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.508880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:86 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.518976] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.518995] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:12157 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.519003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:112 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.528741] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.528760] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:66 nsid:1 lba:3594 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.528768] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:66 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.536533] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.536552] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:21211 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.536560] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:108 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.547173] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.547192] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:84 nsid:1 lba:20595 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.547200] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:84 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.555009] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.555029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:22773 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.555037] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:31 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.565256] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.565276] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:6072 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.565287] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:35 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.575592] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.575610] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:11823 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.575619] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:109 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.583534] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.583553] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:10772 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.583561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:98 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.593595] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.593615] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:20902 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.593622] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:125 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.603263] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.603283] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:24048 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.603290] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:118 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.612047] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.612066] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:476 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.612073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:25 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.621218] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.621242] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:21 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.621250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:26 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.630092] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.630110] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:18541 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.630118] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:92 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.639553] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.639573] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:219 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.639581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:53 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.649723] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.649742] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:24762 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.649750] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:116 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.658150] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.658170] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:4185 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.658178] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:34 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.668652] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.668672] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:2458 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.668679] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:25 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.680404] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.680424] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:23955 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.680432] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:24 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.688550] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.688570] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:110 nsid:1 lba:25100 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.688578] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:110 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:04.976 [2024-07-15 18:11:58.698798] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:04.976 [2024-07-15 18:11:58.698819] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:23729 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:04.976 [2024-07-15 18:11:58.698827] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:126 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:05.234 [2024-07-15 18:11:58.706749] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0x1e7cf20) 00:26:05.234 [2024-07-15 18:11:58.706771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:18663 len:1 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:05.234 [2024-07-15 18:11:58.706779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:26 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:05.234 00:26:05.234 Latency(us) 00:26:05.234 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:05.234 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 128, IO size: 4096) 00:26:05.234 nvme0n1 : 2.04 25832.09 100.91 0.00 0.00 4849.84 2051.56 44906.41 00:26:05.234 =================================================================================================================== 00:26:05.234 Total : 25832.09 100.91 0.00 0.00 4849.84 2051.56 44906.41 00:26:05.234 0 00:26:05.234 18:11:58 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:26:05.234 18:11:58 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:26:05.234 18:11:58 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:26:05.234 | .driver_specific 00:26:05.234 | .nvme_error 00:26:05.234 | .status_code 00:26:05.234 | .command_transient_transport_error' 00:26:05.234 18:11:58 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:26:05.234 18:11:58 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # (( 207 > 0 )) 00:26:05.234 18:11:58 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@73 -- # killprocess 733104 00:26:05.234 18:11:58 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@948 -- # '[' -z 733104 ']' 00:26:05.234 18:11:58 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@952 -- # kill -0 733104 00:26:05.234 18:11:58 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@953 -- # uname 00:26:05.234 18:11:58 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:26:05.234 18:11:58 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 733104 00:26:05.492 18:11:58 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:26:05.492 18:11:58 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:26:05.492 18:11:58 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@966 -- # echo 'killing process with pid 733104' 00:26:05.492 killing process with pid 733104 00:26:05.492 18:11:58 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@967 -- # kill 733104 00:26:05.492 Received shutdown signal, test time was about 2.000000 seconds 00:26:05.492 00:26:05.492 Latency(us) 00:26:05.492 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:05.492 =================================================================================================================== 00:26:05.492 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:26:05.492 18:11:58 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@972 -- # wait 733104 00:26:05.492 18:11:59 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@109 -- # run_bperf_err randread 131072 16 00:26:05.492 18:11:59 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@54 -- # local rw bs qd 00:26:05.492 18:11:59 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # rw=randread 00:26:05.493 18:11:59 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # bs=131072 00:26:05.493 18:11:59 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # qd=16 00:26:05.493 18:11:59 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@58 -- # bperfpid=733804 00:26:05.493 18:11:59 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@60 -- # waitforlisten 733804 /var/tmp/bperf.sock 00:26:05.493 18:11:59 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randread -o 131072 -t 2 -q 16 -z 00:26:05.493 18:11:59 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@829 -- # '[' -z 733804 ']' 00:26:05.493 18:11:59 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bperf.sock 00:26:05.493 18:11:59 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@834 -- # local max_retries=100 00:26:05.493 18:11:59 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:26:05.493 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:26:05.493 18:11:59 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@838 -- # xtrace_disable 00:26:05.493 18:11:59 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:26:05.493 [2024-07-15 18:11:59.216180] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:26:05.493 [2024-07-15 18:11:59.216239] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid733804 ] 00:26:05.493 I/O size of 131072 is greater than zero copy threshold (65536). 00:26:05.493 Zero copy mechanism will not be used. 00:26:05.750 EAL: No free 2048 kB hugepages reported on node 1 00:26:05.750 [2024-07-15 18:11:59.270241] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:05.750 [2024-07-15 18:11:59.348915] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:26:06.316 18:12:00 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:26:06.316 18:12:00 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@862 -- # return 0 00:26:06.316 18:12:00 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:26:06.316 18:12:00 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:26:06.575 18:12:00 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:26:06.575 18:12:00 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:06.575 18:12:00 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:26:06.575 18:12:00 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:06.575 18:12:00 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:26:06.575 18:12:00 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:26:07.144 nvme0n1 00:26:07.144 18:12:00 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 32 00:26:07.144 18:12:00 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:07.144 18:12:00 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:26:07.144 18:12:00 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:07.144 18:12:00 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@69 -- # bperf_py perform_tests 00:26:07.144 18:12:00 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:26:07.144 I/O size of 131072 is greater than zero copy threshold (65536). 00:26:07.144 Zero copy mechanism will not be used. 00:26:07.144 Running I/O for 2 seconds... 00:26:07.144 [2024-07-15 18:12:00.732210] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.144 [2024-07-15 18:12:00.732255] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:15520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.144 [2024-07-15 18:12:00.732266] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.144 [2024-07-15 18:12:00.741058] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.144 [2024-07-15 18:12:00.741082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:19520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.144 [2024-07-15 18:12:00.741091] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.144 [2024-07-15 18:12:00.749466] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.144 [2024-07-15 18:12:00.749487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:20416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.144 [2024-07-15 18:12:00.749500] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.144 [2024-07-15 18:12:00.757169] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.144 [2024-07-15 18:12:00.757190] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:11808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.144 [2024-07-15 18:12:00.757198] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.144 [2024-07-15 18:12:00.764172] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.144 [2024-07-15 18:12:00.764192] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:14944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.144 [2024-07-15 18:12:00.764200] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.144 [2024-07-15 18:12:00.771007] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.144 [2024-07-15 18:12:00.771026] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:21472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.144 [2024-07-15 18:12:00.771034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.144 [2024-07-15 18:12:00.777885] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.144 [2024-07-15 18:12:00.777904] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:11712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.144 [2024-07-15 18:12:00.777912] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.144 [2024-07-15 18:12:00.784667] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.144 [2024-07-15 18:12:00.784687] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:3808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.144 [2024-07-15 18:12:00.784694] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.144 [2024-07-15 18:12:00.791648] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.144 [2024-07-15 18:12:00.791668] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:19424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.144 [2024-07-15 18:12:00.791677] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.145 [2024-07-15 18:12:00.799693] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.145 [2024-07-15 18:12:00.799714] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:15808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.145 [2024-07-15 18:12:00.799722] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.145 [2024-07-15 18:12:00.807216] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.145 [2024-07-15 18:12:00.807241] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:11968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.145 [2024-07-15 18:12:00.807250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.145 [2024-07-15 18:12:00.814070] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.145 [2024-07-15 18:12:00.814089] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:9408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.145 [2024-07-15 18:12:00.814098] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.145 [2024-07-15 18:12:00.821130] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.145 [2024-07-15 18:12:00.821151] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:17696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.145 [2024-07-15 18:12:00.821160] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.145 [2024-07-15 18:12:00.829021] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.145 [2024-07-15 18:12:00.829043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:20800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.145 [2024-07-15 18:12:00.829052] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.145 [2024-07-15 18:12:00.836624] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.145 [2024-07-15 18:12:00.836647] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:13728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.145 [2024-07-15 18:12:00.836657] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.145 [2024-07-15 18:12:00.845218] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.145 [2024-07-15 18:12:00.845257] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:5696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.145 [2024-07-15 18:12:00.845265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.145 [2024-07-15 18:12:00.853902] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.145 [2024-07-15 18:12:00.853925] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:5152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.145 [2024-07-15 18:12:00.853933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.145 [2024-07-15 18:12:00.862725] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.145 [2024-07-15 18:12:00.862748] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:23520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.145 [2024-07-15 18:12:00.862756] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.405 [2024-07-15 18:12:00.872076] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.405 [2024-07-15 18:12:00.872098] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:25472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.405 [2024-07-15 18:12:00.872107] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.405 [2024-07-15 18:12:00.880303] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.405 [2024-07-15 18:12:00.880334] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:23552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.405 [2024-07-15 18:12:00.880346] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.405 [2024-07-15 18:12:00.888820] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.405 [2024-07-15 18:12:00.888842] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:21664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.405 [2024-07-15 18:12:00.888850] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.405 [2024-07-15 18:12:00.895939] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.405 [2024-07-15 18:12:00.895962] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:11552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.405 [2024-07-15 18:12:00.895970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.405 [2024-07-15 18:12:00.902953] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.405 [2024-07-15 18:12:00.902974] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:4928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.405 [2024-07-15 18:12:00.902983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.405 [2024-07-15 18:12:00.909838] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.405 [2024-07-15 18:12:00.909858] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19456 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.405 [2024-07-15 18:12:00.909867] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.405 [2024-07-15 18:12:00.916660] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.405 [2024-07-15 18:12:00.916681] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:20672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.405 [2024-07-15 18:12:00.916689] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.405 [2024-07-15 18:12:00.923512] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.405 [2024-07-15 18:12:00.923533] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:1568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.405 [2024-07-15 18:12:00.923541] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.405 [2024-07-15 18:12:00.930413] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.405 [2024-07-15 18:12:00.930434] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:4544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.405 [2024-07-15 18:12:00.930442] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.405 [2024-07-15 18:12:00.937457] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.405 [2024-07-15 18:12:00.937478] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:6944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.405 [2024-07-15 18:12:00.937486] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.405 [2024-07-15 18:12:00.944198] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.405 [2024-07-15 18:12:00.944222] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:23648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.405 [2024-07-15 18:12:00.944236] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.405 [2024-07-15 18:12:00.950917] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.405 [2024-07-15 18:12:00.950938] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:15072 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.405 [2024-07-15 18:12:00.950947] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.405 [2024-07-15 18:12:00.958318] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.405 [2024-07-15 18:12:00.958339] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:7680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.405 [2024-07-15 18:12:00.958347] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.405 [2024-07-15 18:12:00.966017] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.405 [2024-07-15 18:12:00.966039] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:2976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.405 [2024-07-15 18:12:00.966047] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.405 [2024-07-15 18:12:00.974541] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.405 [2024-07-15 18:12:00.974563] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:1376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.405 [2024-07-15 18:12:00.974571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.405 [2024-07-15 18:12:00.984004] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.405 [2024-07-15 18:12:00.984026] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:14400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.405 [2024-07-15 18:12:00.984034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.405 [2024-07-15 18:12:00.993692] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.405 [2024-07-15 18:12:00.993714] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:10464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.405 [2024-07-15 18:12:00.993723] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.405 [2024-07-15 18:12:01.003401] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.405 [2024-07-15 18:12:01.003423] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:12800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.405 [2024-07-15 18:12:01.003432] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.405 [2024-07-15 18:12:01.013727] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.405 [2024-07-15 18:12:01.013749] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:5088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.406 [2024-07-15 18:12:01.013758] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.406 [2024-07-15 18:12:01.024045] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.406 [2024-07-15 18:12:01.024067] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:18976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.406 [2024-07-15 18:12:01.024075] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.406 [2024-07-15 18:12:01.033488] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.406 [2024-07-15 18:12:01.033509] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:1376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.406 [2024-07-15 18:12:01.033517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.406 [2024-07-15 18:12:01.044587] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.406 [2024-07-15 18:12:01.044609] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.406 [2024-07-15 18:12:01.044617] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.406 [2024-07-15 18:12:01.054368] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.406 [2024-07-15 18:12:01.054389] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:14784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.406 [2024-07-15 18:12:01.054397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.406 [2024-07-15 18:12:01.063724] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.406 [2024-07-15 18:12:01.063745] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:23968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.406 [2024-07-15 18:12:01.063754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.406 [2024-07-15 18:12:01.073359] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.406 [2024-07-15 18:12:01.073381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:5248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.406 [2024-07-15 18:12:01.073390] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.406 [2024-07-15 18:12:01.082857] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.406 [2024-07-15 18:12:01.082878] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:9184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.406 [2024-07-15 18:12:01.082886] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.406 [2024-07-15 18:12:01.092233] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.406 [2024-07-15 18:12:01.092255] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:3776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.406 [2024-07-15 18:12:01.092263] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.406 [2024-07-15 18:12:01.101705] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.406 [2024-07-15 18:12:01.101727] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:22112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.406 [2024-07-15 18:12:01.101739] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.406 [2024-07-15 18:12:01.111285] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.406 [2024-07-15 18:12:01.111308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:19712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.406 [2024-07-15 18:12:01.111317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.406 [2024-07-15 18:12:01.121670] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.406 [2024-07-15 18:12:01.121692] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:8800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.406 [2024-07-15 18:12:01.121700] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.132060] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.132083] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:5376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.132107] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.141782] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.141804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:8608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.141813] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.152400] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.152421] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:3040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.152429] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.161748] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.161771] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:8288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.161779] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.171535] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.171557] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:17504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.171565] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.181005] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.181026] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:11008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.181035] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.191564] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.191589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:17440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.191597] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.202351] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.202372] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:14432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.202380] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.211558] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.211580] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:704 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.211588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.221669] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.221689] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:19392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.221697] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.231306] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.231328] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:22176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.231337] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.241223] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.241250] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:5888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.241259] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.251724] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.251745] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:11328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.251752] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.260955] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.260976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.260984] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.270212] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.270239] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:22144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.270247] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.279339] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.279361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:12896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.279369] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.288973] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.288995] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:20096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.289003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.298340] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.298361] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:23488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.298370] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.307230] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.307252] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:2176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.307260] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.316628] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.316650] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:14112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.316658] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.326939] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.326961] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:6624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.326969] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.337098] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.337120] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:12224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.337128] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.348114] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.348135] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:11136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.348143] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.359060] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.666 [2024-07-15 18:12:01.359081] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:24480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.666 [2024-07-15 18:12:01.359094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.666 [2024-07-15 18:12:01.367969] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.667 [2024-07-15 18:12:01.367991] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:9184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.667 [2024-07-15 18:12:01.367999] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.667 [2024-07-15 18:12:01.376803] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.667 [2024-07-15 18:12:01.376825] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:2304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.667 [2024-07-15 18:12:01.376833] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.667 [2024-07-15 18:12:01.385713] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.667 [2024-07-15 18:12:01.385734] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:3200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.667 [2024-07-15 18:12:01.385742] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.926 [2024-07-15 18:12:01.395122] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.926 [2024-07-15 18:12:01.395144] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:8256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.926 [2024-07-15 18:12:01.395153] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.926 [2024-07-15 18:12:01.403913] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.926 [2024-07-15 18:12:01.403935] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:6336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.926 [2024-07-15 18:12:01.403944] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.414345] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.414367] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:18112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.414376] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.423641] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.423662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:24224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.423671] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.433443] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.433465] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:9152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.433474] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.442986] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.443012] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:22080 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.443020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.453009] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.453030] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:12992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.453038] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.462970] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.462991] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:21152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.462999] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.472517] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.472539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.472547] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.483096] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.483118] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:22624 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.483127] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.493092] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.493113] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:9312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.493121] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.503679] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.503700] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:17280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.503709] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.513098] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.513119] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:13856 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.513127] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.523338] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.523359] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:23840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.523367] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.533625] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.533647] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:20672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.533655] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.542886] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.542908] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:1536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.542917] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.552416] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.552438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:10272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.552446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.561835] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.561860] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:21600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.561868] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.571366] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.571389] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:17376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.571398] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.580484] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.580507] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:4448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.580516] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.589929] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.589951] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:16000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.589959] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.599458] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.599480] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:13312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.599489] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.608949] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.608971] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:21152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.608986] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.618393] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.618415] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:5728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.618423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.627607] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.627629] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:7584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.627636] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.636614] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.636636] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:13312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.636644] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:07.927 [2024-07-15 18:12:01.646100] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:07.927 [2024-07-15 18:12:01.646122] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:3264 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:07.927 [2024-07-15 18:12:01.646131] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.186 [2024-07-15 18:12:01.655138] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.186 [2024-07-15 18:12:01.655160] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:4096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.186 [2024-07-15 18:12:01.655169] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.186 [2024-07-15 18:12:01.664886] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.186 [2024-07-15 18:12:01.664909] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:0 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.186 [2024-07-15 18:12:01.664917] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.186 [2024-07-15 18:12:01.674097] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.186 [2024-07-15 18:12:01.674119] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:11904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.186 [2024-07-15 18:12:01.674128] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.186 [2024-07-15 18:12:01.683675] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.186 [2024-07-15 18:12:01.683712] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:9120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.186 [2024-07-15 18:12:01.683721] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.186 [2024-07-15 18:12:01.692865] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.186 [2024-07-15 18:12:01.692891] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:23328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.186 [2024-07-15 18:12:01.692900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.186 [2024-07-15 18:12:01.701025] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.186 [2024-07-15 18:12:01.701047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:1568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.186 [2024-07-15 18:12:01.701055] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.186 [2024-07-15 18:12:01.709898] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.186 [2024-07-15 18:12:01.709920] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:4864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.709928] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.719022] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.719044] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:22176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.719053] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.727265] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.727286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:25280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.727295] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.736167] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.736190] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.736198] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.744432] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.744454] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.744463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.752369] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.752391] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:14176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.752399] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.759206] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.759232] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:14272 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.759241] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.765640] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.765662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.765669] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.771981] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.772002] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:11456 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.772010] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.778265] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.778285] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:12544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.778293] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.784558] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.784579] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:15648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.784587] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.790852] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.790872] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:12608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.790880] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.796882] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.796902] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:24544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.796911] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.803405] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.803426] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:20800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.803434] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.810255] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.810277] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:13280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.810285] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.818258] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.818280] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:1664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.818291] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.826835] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.826857] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:21728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.826865] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.835852] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.835873] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:4928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.835882] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.844338] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.844360] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:17088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.844368] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.852712] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.852733] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:7136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.852742] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.861237] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.861259] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:3552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.187 [2024-07-15 18:12:01.861267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.187 [2024-07-15 18:12:01.869752] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.187 [2024-07-15 18:12:01.869774] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:3264 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.188 [2024-07-15 18:12:01.869782] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.188 [2024-07-15 18:12:01.879256] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.188 [2024-07-15 18:12:01.879278] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:6848 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.188 [2024-07-15 18:12:01.879287] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.188 [2024-07-15 18:12:01.888272] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.188 [2024-07-15 18:12:01.888294] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:22912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.188 [2024-07-15 18:12:01.888303] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.188 [2024-07-15 18:12:01.897477] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.188 [2024-07-15 18:12:01.897498] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:18112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.188 [2024-07-15 18:12:01.897506] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.188 [2024-07-15 18:12:01.906092] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.188 [2024-07-15 18:12:01.906114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:320 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.188 [2024-07-15 18:12:01.906122] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.448 [2024-07-15 18:12:01.915044] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.448 [2024-07-15 18:12:01.915068] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:21216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.448 [2024-07-15 18:12:01.915077] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.448 [2024-07-15 18:12:01.922152] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.448 [2024-07-15 18:12:01.922174] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:14784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.448 [2024-07-15 18:12:01.922182] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.448 [2024-07-15 18:12:01.930347] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.448 [2024-07-15 18:12:01.930369] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.448 [2024-07-15 18:12:01.930378] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.448 [2024-07-15 18:12:01.939329] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.448 [2024-07-15 18:12:01.939352] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:23968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.448 [2024-07-15 18:12:01.939360] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.448 [2024-07-15 18:12:01.948771] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.448 [2024-07-15 18:12:01.948795] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:18432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.448 [2024-07-15 18:12:01.948805] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.448 [2024-07-15 18:12:01.958733] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.448 [2024-07-15 18:12:01.958755] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:4448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.448 [2024-07-15 18:12:01.958764] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.448 [2024-07-15 18:12:01.969404] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.448 [2024-07-15 18:12:01.969425] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:15840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.448 [2024-07-15 18:12:01.969437] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.448 [2024-07-15 18:12:01.978936] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.448 [2024-07-15 18:12:01.978958] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:7648 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.448 [2024-07-15 18:12:01.978966] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.448 [2024-07-15 18:12:01.988512] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.448 [2024-07-15 18:12:01.988534] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:8512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.448 [2024-07-15 18:12:01.988542] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.448 [2024-07-15 18:12:01.997016] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.448 [2024-07-15 18:12:01.997038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:19104 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.448 [2024-07-15 18:12:01.997046] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.448 [2024-07-15 18:12:02.005231] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.005253] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:5632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.005262] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.013056] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.013078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:10048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.013087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.020361] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.020383] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:4928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.020392] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.027836] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.027857] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:23008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.027866] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.035177] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.035198] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:21952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.035207] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.042117] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.042141] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:2336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.042149] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.048677] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.048698] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:20800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.048709] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.055511] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.055532] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:22976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.055541] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.061956] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.061977] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:1792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.061985] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.070854] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.070875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:23808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.070883] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.078872] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.078892] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:6112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.078900] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.086222] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.086248] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:4736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.086256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.093509] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.093530] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:24032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.093538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.100127] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.100148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:1600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.100156] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.106965] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.106986] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:20064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.106994] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.114124] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.114146] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:11936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.114154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.121733] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.121755] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:18432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.121764] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.130480] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.130503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:12928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.130511] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.138865] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.138886] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:11616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.138894] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.147014] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.147034] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:10912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.147042] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.155067] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.155087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:14784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.155095] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.163049] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.163071] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:5344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.163079] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.449 [2024-07-15 18:12:02.170697] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.449 [2024-07-15 18:12:02.170719] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:20512 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.449 [2024-07-15 18:12:02.170731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.708 [2024-07-15 18:12:02.178968] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.708 [2024-07-15 18:12:02.178990] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:18656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.708 [2024-07-15 18:12:02.178999] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.708 [2024-07-15 18:12:02.187445] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.708 [2024-07-15 18:12:02.187465] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.708 [2024-07-15 18:12:02.187473] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.708 [2024-07-15 18:12:02.195699] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.708 [2024-07-15 18:12:02.195720] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:2848 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.708 [2024-07-15 18:12:02.195727] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.708 [2024-07-15 18:12:02.203313] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.708 [2024-07-15 18:12:02.203335] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:1888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.708 [2024-07-15 18:12:02.203343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.708 [2024-07-15 18:12:02.211716] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.708 [2024-07-15 18:12:02.211737] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:22016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.211745] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.218877] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.218897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:4352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.218905] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:4 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.225679] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.225699] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:22656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.225707] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.229253] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.229273] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:19968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.229281] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.236842] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.236866] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:13344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.236874] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.243288] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.243308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:1088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.243316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.250133] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.250153] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:22144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.250161] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.256870] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.256891] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:18720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.256899] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.263838] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.263859] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:19392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.263867] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.271471] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.271492] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:1760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.271500] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.278662] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.278683] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:4224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.278691] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.286599] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.286620] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:9120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.286628] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.296721] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.296741] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:8064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.296749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.306582] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.306603] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:9856 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.306610] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.316301] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.316321] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:24576 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.316330] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.325436] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.325456] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:2656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.325464] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.334848] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.334870] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:13248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.334878] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.344995] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.345015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:13088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.345023] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.355015] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.355036] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:24736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.355044] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.363833] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.363853] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:18528 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.363861] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.371721] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.371742] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:20192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.371749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.378976] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.378997] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:2720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.379009] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.386180] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.386201] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:17248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.386209] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.393059] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.393079] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:6752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.393087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.399878] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.399898] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:13664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.399905] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.406384] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.406404] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:20768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.406412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.412488] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.412507] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:8224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.412515] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.418650] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.418671] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:5024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.418679] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.424344] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.424364] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:23936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.709 [2024-07-15 18:12:02.424372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.709 [2024-07-15 18:12:02.430567] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.709 [2024-07-15 18:12:02.430588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:14432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.710 [2024-07-15 18:12:02.430595] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.969 [2024-07-15 18:12:02.436874] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.969 [2024-07-15 18:12:02.436900] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.969 [2024-07-15 18:12:02.436908] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.969 [2024-07-15 18:12:02.443287] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.969 [2024-07-15 18:12:02.443307] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:16608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.969 [2024-07-15 18:12:02.443315] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.969 [2024-07-15 18:12:02.449430] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.969 [2024-07-15 18:12:02.449450] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:10240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.969 [2024-07-15 18:12:02.449458] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.969 [2024-07-15 18:12:02.455290] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.969 [2024-07-15 18:12:02.455310] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:3200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.969 [2024-07-15 18:12:02.455317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.969 [2024-07-15 18:12:02.462730] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.969 [2024-07-15 18:12:02.462750] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:1664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.969 [2024-07-15 18:12:02.462758] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.969 [2024-07-15 18:12:02.471847] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.969 [2024-07-15 18:12:02.471867] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:9504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.969 [2024-07-15 18:12:02.471876] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.969 [2024-07-15 18:12:02.480676] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.969 [2024-07-15 18:12:02.480696] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:2816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.969 [2024-07-15 18:12:02.480704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.969 [2024-07-15 18:12:02.488352] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.969 [2024-07-15 18:12:02.488372] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:4896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.969 [2024-07-15 18:12:02.488380] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.969 [2024-07-15 18:12:02.495608] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.969 [2024-07-15 18:12:02.495629] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:6464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.969 [2024-07-15 18:12:02.495637] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.969 [2024-07-15 18:12:02.503058] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.969 [2024-07-15 18:12:02.503079] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:1376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.969 [2024-07-15 18:12:02.503087] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.969 [2024-07-15 18:12:02.509485] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.969 [2024-07-15 18:12:02.509505] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:6944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.969 [2024-07-15 18:12:02.509513] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.969 [2024-07-15 18:12:02.515448] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.969 [2024-07-15 18:12:02.515469] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:7296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.969 [2024-07-15 18:12:02.515476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.521954] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.521975] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:10080 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.521982] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.527855] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.527875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:11360 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.527883] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.533860] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.533880] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:15072 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.533889] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.539548] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.539568] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:2656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.539576] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.545473] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.545493] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:23200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.545501] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.551795] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.551816] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:12064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.551826] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.557506] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.557527] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:16992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.557535] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.563200] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.563220] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:11488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.563234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.569319] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.569339] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:10816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.569346] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.577615] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.577635] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:9408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.577643] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.586529] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.586549] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:7296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.586557] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.594560] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.594580] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:15808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.594588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:7 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.602359] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.602379] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:22464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.602386] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.609867] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.609887] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:20992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.609895] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:5 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.618953] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.618979] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:24896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.618987] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.627156] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.627175] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:7424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.627183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.634605] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.634625] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:13984 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.634633] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.641781] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.641801] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:5568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.641809] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.645240] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.645260] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:5984 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.645267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.652081] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.652101] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:24672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.652109] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.658135] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.658155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:22560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.658163] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.663798] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.663818] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:7424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.663826] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.670101] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.670122] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:20288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.670130] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.675168] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.675189] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:22592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.675196] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.681031] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.681051] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:10208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.681059] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.686612] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.686632] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:2752 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.686640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:08.970 [2024-07-15 18:12:02.692067] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:08.970 [2024-07-15 18:12:02.692087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:16192 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:08.970 [2024-07-15 18:12:02.692095] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:09.230 [2024-07-15 18:12:02.697452] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:09.230 [2024-07-15 18:12:02.697473] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:9824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:09.230 [2024-07-15 18:12:02.697482] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:09.230 [2024-07-15 18:12:02.703287] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:09.230 [2024-07-15 18:12:02.703308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:5952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:09.230 [2024-07-15 18:12:02.703316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:09.230 [2024-07-15 18:12:02.708394] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:09.230 [2024-07-15 18:12:02.708414] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:6144 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:09.230 [2024-07-15 18:12:02.708423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:6 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:09.230 [2024-07-15 18:12:02.714387] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:09.230 [2024-07-15 18:12:02.714407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:18400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:09.230 [2024-07-15 18:12:02.714415] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:09.230 [2024-07-15 18:12:02.723541] nvme_tcp.c:1459:nvme_tcp_accel_seq_recv_compute_crc32_done: *ERROR*: data digest error on tqpair=(0xd4e0b0) 00:26:09.230 [2024-07-15 18:12:02.723562] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:19040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:09.230 [2024-07-15 18:12:02.723573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:09.230 00:26:09.230 Latency(us) 00:26:09.230 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:09.230 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 16, IO size: 131072) 00:26:09.230 nvme0n1 : 2.00 3803.39 475.42 0.00 0.00 4202.84 701.66 11739.49 00:26:09.230 =================================================================================================================== 00:26:09.230 Total : 3803.39 475.42 0.00 0.00 4202.84 701.66 11739.49 00:26:09.230 0 00:26:09.230 18:12:02 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:26:09.230 18:12:02 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:26:09.230 18:12:02 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:26:09.230 | .driver_specific 00:26:09.230 | .nvme_error 00:26:09.230 | .status_code 00:26:09.230 | .command_transient_transport_error' 00:26:09.230 18:12:02 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:26:09.230 18:12:02 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # (( 245 > 0 )) 00:26:09.230 18:12:02 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@73 -- # killprocess 733804 00:26:09.230 18:12:02 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@948 -- # '[' -z 733804 ']' 00:26:09.230 18:12:02 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@952 -- # kill -0 733804 00:26:09.230 18:12:02 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@953 -- # uname 00:26:09.230 18:12:02 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:26:09.230 18:12:02 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 733804 00:26:09.489 18:12:02 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:26:09.489 18:12:02 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:26:09.489 18:12:02 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@966 -- # echo 'killing process with pid 733804' 00:26:09.489 killing process with pid 733804 00:26:09.489 18:12:02 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@967 -- # kill 733804 00:26:09.489 Received shutdown signal, test time was about 2.000000 seconds 00:26:09.489 00:26:09.489 Latency(us) 00:26:09.489 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:09.489 =================================================================================================================== 00:26:09.489 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:26:09.489 18:12:02 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@972 -- # wait 733804 00:26:09.489 18:12:03 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@114 -- # run_bperf_err randwrite 4096 128 00:26:09.489 18:12:03 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@54 -- # local rw bs qd 00:26:09.489 18:12:03 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # rw=randwrite 00:26:09.489 18:12:03 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # bs=4096 00:26:09.489 18:12:03 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # qd=128 00:26:09.489 18:12:03 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@58 -- # bperfpid=734510 00:26:09.489 18:12:03 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@60 -- # waitforlisten 734510 /var/tmp/bperf.sock 00:26:09.489 18:12:03 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 4096 -t 2 -q 128 -z 00:26:09.489 18:12:03 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@829 -- # '[' -z 734510 ']' 00:26:09.489 18:12:03 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bperf.sock 00:26:09.489 18:12:03 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@834 -- # local max_retries=100 00:26:09.489 18:12:03 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:26:09.489 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:26:09.489 18:12:03 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@838 -- # xtrace_disable 00:26:09.489 18:12:03 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:26:09.489 [2024-07-15 18:12:03.198911] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:26:09.489 [2024-07-15 18:12:03.198964] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid734510 ] 00:26:09.747 EAL: No free 2048 kB hugepages reported on node 1 00:26:09.747 [2024-07-15 18:12:03.253023] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:09.747 [2024-07-15 18:12:03.332364] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:26:10.313 18:12:04 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:26:10.313 18:12:04 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@862 -- # return 0 00:26:10.313 18:12:04 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:26:10.313 18:12:04 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:26:10.573 18:12:04 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:26:10.573 18:12:04 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:10.573 18:12:04 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:26:10.573 18:12:04 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:10.573 18:12:04 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:26:10.573 18:12:04 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:26:11.141 nvme0n1 00:26:11.141 18:12:04 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 256 00:26:11.141 18:12:04 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:11.141 18:12:04 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:26:11.141 18:12:04 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:11.141 18:12:04 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@69 -- # bperf_py perform_tests 00:26:11.141 18:12:04 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:26:11.141 Running I/O for 2 seconds... 00:26:11.141 [2024-07-15 18:12:04.706172] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f9f68 00:26:11.141 [2024-07-15 18:12:04.707050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:13017 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.141 [2024-07-15 18:12:04.707085] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:81 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:11.141 [2024-07-15 18:12:04.716544] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190df118 00:26:11.141 [2024-07-15 18:12:04.717646] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:81 nsid:1 lba:13626 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.141 [2024-07-15 18:12:04.717670] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:81 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:26:11.141 [2024-07-15 18:12:04.725874] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e01f8 00:26:11.141 [2024-07-15 18:12:04.726948] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:78 nsid:1 lba:4828 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.141 [2024-07-15 18:12:04.726968] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:78 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:26:11.141 [2024-07-15 18:12:04.735195] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190feb58 00:26:11.141 [2024-07-15 18:12:04.736236] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:3 nsid:1 lba:10902 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.141 [2024-07-15 18:12:04.736257] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:3 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:26:11.141 [2024-07-15 18:12:04.744362] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190fe2e8 00:26:11.141 [2024-07-15 18:12:04.745429] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:62 nsid:1 lba:13003 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.141 [2024-07-15 18:12:04.745449] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:62 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:26:11.141 [2024-07-15 18:12:04.753536] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f4b08 00:26:11.141 [2024-07-15 18:12:04.754611] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:77 nsid:1 lba:6783 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.141 [2024-07-15 18:12:04.754630] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:77 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:26:11.141 [2024-07-15 18:12:04.762689] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f3a28 00:26:11.141 [2024-07-15 18:12:04.763651] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:76 nsid:1 lba:20639 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.141 [2024-07-15 18:12:04.763670] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:76 cdw0:0 sqhd:0070 p:0 m:0 dnr:0 00:26:11.141 [2024-07-15 18:12:04.771262] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f6020 00:26:11.141 [2024-07-15 18:12:04.772244] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:75 nsid:1 lba:5584 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.141 [2024-07-15 18:12:04.772264] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:75 cdw0:0 sqhd:0031 p:0 m:0 dnr:0 00:26:11.141 [2024-07-15 18:12:04.780816] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e88f8 00:26:11.141 [2024-07-15 18:12:04.781989] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:24628 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.141 [2024-07-15 18:12:04.782008] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:115 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:11.141 [2024-07-15 18:12:04.791800] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190df118 00:26:11.141 [2024-07-15 18:12:04.793349] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:3946 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.141 [2024-07-15 18:12:04.793368] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:11.141 [2024-07-15 18:12:04.798185] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190ed920 00:26:11.141 [2024-07-15 18:12:04.798994] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:14 nsid:1 lba:19017 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.141 [2024-07-15 18:12:04.799014] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:14 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:26:11.141 [2024-07-15 18:12:04.807363] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f20d8 00:26:11.141 [2024-07-15 18:12:04.808147] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:2720 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.141 [2024-07-15 18:12:04.808166] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:003f p:0 m:0 dnr:0 00:26:11.141 [2024-07-15 18:12:04.816610] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190ff3c8 00:26:11.141 [2024-07-15 18:12:04.817404] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:25187 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.141 [2024-07-15 18:12:04.817423] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.141 [2024-07-15 18:12:04.825660] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f57b0 00:26:11.141 [2024-07-15 18:12:04.826450] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:11431 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.141 [2024-07-15 18:12:04.826468] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:86 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.141 [2024-07-15 18:12:04.834803] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f46d0 00:26:11.141 [2024-07-15 18:12:04.835589] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:17348 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.141 [2024-07-15 18:12:04.835607] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:72 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.141 [2024-07-15 18:12:04.843889] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f2948 00:26:11.141 [2024-07-15 18:12:04.844675] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:1288 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.141 [2024-07-15 18:12:04.844693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:25 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.141 [2024-07-15 18:12:04.852994] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f96f8 00:26:11.141 [2024-07-15 18:12:04.853769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:20824 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.141 [2024-07-15 18:12:04.853788] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:24 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.141 [2024-07-15 18:12:04.862155] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f8618 00:26:11.141 [2024-07-15 18:12:04.862967] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:7441 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.141 [2024-07-15 18:12:04.862986] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:23 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:04.871385] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f7538 00:26:11.400 [2024-07-15 18:12:04.872172] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:2214 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:04.872191] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:22 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:04.880493] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f6458 00:26:11.400 [2024-07-15 18:12:04.881307] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:21054 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:04.881327] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:21 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:04.889626] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e4578 00:26:11.400 [2024-07-15 18:12:04.890341] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:16321 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:04.890360] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:20 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:04.898664] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e5658 00:26:11.400 [2024-07-15 18:12:04.899462] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:8115 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:04.899481] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:19 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:04.907925] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e6738 00:26:11.400 [2024-07-15 18:12:04.908718] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:10807 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:04.908737] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:18 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:04.917039] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190ed920 00:26:11.400 [2024-07-15 18:12:04.917857] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:6819 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:04.917876] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:17 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:04.926346] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190eea00 00:26:11.400 [2024-07-15 18:12:04.927131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:19711 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:04.927150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:16 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:04.935607] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190efae0 00:26:11.400 [2024-07-15 18:12:04.936411] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4553 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:04.936429] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:04.944650] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f0bc0 00:26:11.400 [2024-07-15 18:12:04.945429] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:6510 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:04.945451] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:79 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:04.953750] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f1ca0 00:26:11.400 [2024-07-15 18:12:04.954546] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:9165 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:04.954565] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:80 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:04.962982] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190feb58 00:26:11.400 [2024-07-15 18:12:04.963790] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:23329 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:04.963810] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:82 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:04.972276] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190fe2e8 00:26:11.400 [2024-07-15 18:12:04.973082] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:15550 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:04.973100] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:83 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:04.981555] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f4b08 00:26:11.400 [2024-07-15 18:12:04.982358] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:17651 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:04.982377] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:84 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:04.990827] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f3a28 00:26:11.400 [2024-07-15 18:12:04.991623] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:5758 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:04.991642] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:73 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:05.000047] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190fd640 00:26:11.400 [2024-07-15 18:12:05.000826] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:1255 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:05.000845] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:33 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:05.009182] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f92c0 00:26:11.400 [2024-07-15 18:12:05.009978] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:20526 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:05.009996] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:34 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:05.018302] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f81e0 00:26:11.400 [2024-07-15 18:12:05.019083] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:18205 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:05.019101] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:35 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:05.027454] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f7100 00:26:11.400 [2024-07-15 18:12:05.028238] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:5288 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:05.028259] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:36 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:05.036445] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e38d0 00:26:11.400 [2024-07-15 18:12:05.037232] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:18368 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:05.037251] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:37 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:05.045520] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e49b0 00:26:11.400 [2024-07-15 18:12:05.046302] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:4069 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:05.046321] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:38 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:05.054654] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e5a90 00:26:11.400 [2024-07-15 18:12:05.055438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:13301 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:05.055456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:05.063732] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f5be8 00:26:11.400 [2024-07-15 18:12:05.064511] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:24434 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:05.064530] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:40 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:05.072865] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190edd58 00:26:11.400 [2024-07-15 18:12:05.073639] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:24011 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:05.073657] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:41 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:05.081906] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190eee38 00:26:11.400 [2024-07-15 18:12:05.082726] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:8670 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:05.082745] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:42 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:05.091211] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190eff18 00:26:11.400 [2024-07-15 18:12:05.091984] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:13104 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:05.092002] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:43 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:05.100218] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f0ff8 00:26:11.400 [2024-07-15 18:12:05.101031] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:17709 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:05.101049] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:44 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:05.109383] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f20d8 00:26:11.400 [2024-07-15 18:12:05.110155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:9176 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:05.110174] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:45 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.400 [2024-07-15 18:12:05.118460] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190ff3c8 00:26:11.400 [2024-07-15 18:12:05.119237] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:14404 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.400 [2024-07-15 18:12:05.119255] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:26 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.127637] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f57b0 00:26:11.659 [2024-07-15 18:12:05.128436] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:10807 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.128456] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.136821] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f46d0 00:26:11.659 [2024-07-15 18:12:05.137628] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:3933 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.137647] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:101 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.145858] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f2948 00:26:11.659 [2024-07-15 18:12:05.146655] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:24423 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.146673] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.155083] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f96f8 00:26:11.659 [2024-07-15 18:12:05.155862] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:883 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.155881] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:46 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.164147] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f8618 00:26:11.659 [2024-07-15 18:12:05.164924] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:12833 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.164944] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:85 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.173196] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f7538 00:26:11.659 [2024-07-15 18:12:05.173972] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:22709 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.173991] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:115 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.182340] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f6458 00:26:11.659 [2024-07-15 18:12:05.183043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:2215 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.183062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:27 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.191388] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e4578 00:26:11.659 [2024-07-15 18:12:05.192093] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:21494 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.192112] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:28 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.200434] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e5658 00:26:11.659 [2024-07-15 18:12:05.201133] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:23991 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.201151] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:29 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.209559] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e6738 00:26:11.659 [2024-07-15 18:12:05.210293] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:12058 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.210312] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:30 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.218649] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190ed920 00:26:11.659 [2024-07-15 18:12:05.219384] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:15479 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.219404] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:31 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.227906] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190eea00 00:26:11.659 [2024-07-15 18:12:05.228637] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:4233 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.228656] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.237163] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190efae0 00:26:11.659 [2024-07-15 18:12:05.237883] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:11118 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.237902] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.246397] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f0bc0 00:26:11.659 [2024-07-15 18:12:05.247119] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:1440 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.247138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.255521] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f1ca0 00:26:11.659 [2024-07-15 18:12:05.256241] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:21096 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.256260] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.264576] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190feb58 00:26:11.659 [2024-07-15 18:12:05.265299] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:19481 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.265323] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.273631] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190fe2e8 00:26:11.659 [2024-07-15 18:12:05.274336] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:10095 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.274355] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.282742] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f4b08 00:26:11.659 [2024-07-15 18:12:05.283450] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:21127 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.283468] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.291789] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f3a28 00:26:11.659 [2024-07-15 18:12:05.292491] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:2744 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.292509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.300885] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190fd640 00:26:11.659 [2024-07-15 18:12:05.301592] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:25534 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.301611] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.309929] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f92c0 00:26:11.659 [2024-07-15 18:12:05.310634] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:11666 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.310653] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:86 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.318981] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f81e0 00:26:11.659 [2024-07-15 18:12:05.319711] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:14661 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.319731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:72 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.328098] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f7100 00:26:11.659 [2024-07-15 18:12:05.328804] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:8278 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.328823] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:25 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.337222] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e38d0 00:26:11.659 [2024-07-15 18:12:05.337925] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:23751 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.337944] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:24 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.346267] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e49b0 00:26:11.659 [2024-07-15 18:12:05.346968] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:5981 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.346987] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:23 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.355367] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e5a90 00:26:11.659 [2024-07-15 18:12:05.356064] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:6270 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.356083] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:22 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.364525] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f5be8 00:26:11.659 [2024-07-15 18:12:05.365219] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:18148 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.365241] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:21 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.373702] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190edd58 00:26:11.659 [2024-07-15 18:12:05.374409] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:20141 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.374428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:20 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.659 [2024-07-15 18:12:05.382778] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190eee38 00:26:11.659 [2024-07-15 18:12:05.383501] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:16905 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.659 [2024-07-15 18:12:05.383520] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:19 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.918 [2024-07-15 18:12:05.392019] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190eff18 00:26:11.918 [2024-07-15 18:12:05.392756] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:12293 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.918 [2024-07-15 18:12:05.392775] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:18 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.918 [2024-07-15 18:12:05.401302] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f0ff8 00:26:11.918 [2024-07-15 18:12:05.401997] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:5296 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.918 [2024-07-15 18:12:05.402016] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:17 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.918 [2024-07-15 18:12:05.410401] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f20d8 00:26:11.918 [2024-07-15 18:12:05.411108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:7452 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.918 [2024-07-15 18:12:05.411127] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:16 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.918 [2024-07-15 18:12:05.419635] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190ff3c8 00:26:11.918 [2024-07-15 18:12:05.420343] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7292 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.918 [2024-07-15 18:12:05.420362] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.918 [2024-07-15 18:12:05.428771] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f57b0 00:26:11.918 [2024-07-15 18:12:05.429495] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:23260 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.918 [2024-07-15 18:12:05.429514] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:79 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.918 [2024-07-15 18:12:05.437862] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f46d0 00:26:11.918 [2024-07-15 18:12:05.438566] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:16570 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.918 [2024-07-15 18:12:05.438585] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:80 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.918 [2024-07-15 18:12:05.446833] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f2948 00:26:11.918 [2024-07-15 18:12:05.447561] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:24048 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.918 [2024-07-15 18:12:05.447581] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:82 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.918 [2024-07-15 18:12:05.456090] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f96f8 00:26:11.918 [2024-07-15 18:12:05.456812] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:7503 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.918 [2024-07-15 18:12:05.456831] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:83 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.918 [2024-07-15 18:12:05.465287] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f8618 00:26:11.918 [2024-07-15 18:12:05.465986] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:23864 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.918 [2024-07-15 18:12:05.466005] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:84 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.919 [2024-07-15 18:12:05.474342] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f7538 00:26:11.919 [2024-07-15 18:12:05.475064] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:9392 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.919 [2024-07-15 18:12:05.475083] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:73 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.919 [2024-07-15 18:12:05.483588] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f6458 00:26:11.919 [2024-07-15 18:12:05.484315] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:12085 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.919 [2024-07-15 18:12:05.484334] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:33 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.919 [2024-07-15 18:12:05.492830] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e4578 00:26:11.919 [2024-07-15 18:12:05.493557] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:18633 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.919 [2024-07-15 18:12:05.493576] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:34 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.919 [2024-07-15 18:12:05.501887] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e5658 00:26:11.919 [2024-07-15 18:12:05.502591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:10876 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.919 [2024-07-15 18:12:05.502614] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:35 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.919 [2024-07-15 18:12:05.510996] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e6738 00:26:11.919 [2024-07-15 18:12:05.511697] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:21235 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.919 [2024-07-15 18:12:05.511716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:36 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.919 [2024-07-15 18:12:05.520023] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190ed920 00:26:11.919 [2024-07-15 18:12:05.520730] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:9510 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.919 [2024-07-15 18:12:05.520749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:37 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.919 [2024-07-15 18:12:05.529061] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190eea00 00:26:11.919 [2024-07-15 18:12:05.529770] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:24288 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.919 [2024-07-15 18:12:05.529789] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:38 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.919 [2024-07-15 18:12:05.538192] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190efae0 00:26:11.919 [2024-07-15 18:12:05.538901] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:6973 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.919 [2024-07-15 18:12:05.538920] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.919 [2024-07-15 18:12:05.547238] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f0bc0 00:26:11.919 [2024-07-15 18:12:05.547978] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:23697 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.919 [2024-07-15 18:12:05.547998] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:40 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.919 [2024-07-15 18:12:05.556510] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f1ca0 00:26:11.919 [2024-07-15 18:12:05.557236] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:7214 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.919 [2024-07-15 18:12:05.557255] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:41 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.919 [2024-07-15 18:12:05.565609] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190feb58 00:26:11.919 [2024-07-15 18:12:05.566314] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:7486 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.919 [2024-07-15 18:12:05.566332] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:42 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.919 [2024-07-15 18:12:05.574819] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190fe2e8 00:26:11.919 [2024-07-15 18:12:05.575550] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:21308 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.919 [2024-07-15 18:12:05.575569] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:43 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.919 [2024-07-15 18:12:05.584082] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f4b08 00:26:11.919 [2024-07-15 18:12:05.584807] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:9036 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.919 [2024-07-15 18:12:05.584826] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:44 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.919 [2024-07-15 18:12:05.593322] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f3a28 00:26:11.919 [2024-07-15 18:12:05.594042] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:15081 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.919 [2024-07-15 18:12:05.594061] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:45 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.919 [2024-07-15 18:12:05.602422] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190fd640 00:26:11.919 [2024-07-15 18:12:05.603124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:24890 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.919 [2024-07-15 18:12:05.603143] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:26 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.919 [2024-07-15 18:12:05.611609] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f92c0 00:26:11.919 [2024-07-15 18:12:05.612336] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:13166 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.919 [2024-07-15 18:12:05.612355] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.919 [2024-07-15 18:12:05.620726] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f81e0 00:26:11.919 [2024-07-15 18:12:05.621430] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:10130 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.919 [2024-07-15 18:12:05.621450] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:101 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.919 [2024-07-15 18:12:05.629812] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f7100 00:26:11.919 [2024-07-15 18:12:05.630531] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:3106 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.919 [2024-07-15 18:12:05.630550] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:11.919 [2024-07-15 18:12:05.638849] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e38d0 00:26:11.919 [2024-07-15 18:12:05.639551] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:3545 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:11.919 [2024-07-15 18:12:05.639571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:46 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.178 [2024-07-15 18:12:05.648089] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e49b0 00:26:12.178 [2024-07-15 18:12:05.648813] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:13301 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.178 [2024-07-15 18:12:05.648832] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:85 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.178 [2024-07-15 18:12:05.657259] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e5a90 00:26:12.178 [2024-07-15 18:12:05.657959] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:25217 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.178 [2024-07-15 18:12:05.657978] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:115 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.178 [2024-07-15 18:12:05.666292] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f5be8 00:26:12.178 [2024-07-15 18:12:05.666996] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:16598 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.667015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:27 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.675329] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190edd58 00:26:12.179 [2024-07-15 18:12:05.676031] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:15830 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.676049] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:28 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.684482] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190eee38 00:26:12.179 [2024-07-15 18:12:05.685180] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:8743 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.685199] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:29 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.693514] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190eff18 00:26:12.179 [2024-07-15 18:12:05.694210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:17387 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.694233] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:30 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.702550] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f0ff8 00:26:12.179 [2024-07-15 18:12:05.703247] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:11823 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.703267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:31 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.711686] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f20d8 00:26:12.179 [2024-07-15 18:12:05.712407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:8597 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.712426] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.720925] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190ff3c8 00:26:12.179 [2024-07-15 18:12:05.721659] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:24198 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.721679] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.730070] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f57b0 00:26:12.179 [2024-07-15 18:12:05.730811] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:4175 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.730830] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.739510] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f46d0 00:26:12.179 [2024-07-15 18:12:05.740242] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:905 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.740265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.748743] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f2948 00:26:12.179 [2024-07-15 18:12:05.749469] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:1328 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.749488] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.757988] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f96f8 00:26:12.179 [2024-07-15 18:12:05.758712] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:3782 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.758731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.767068] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f8618 00:26:12.179 [2024-07-15 18:12:05.767773] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:3803 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.767792] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.776099] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f7538 00:26:12.179 [2024-07-15 18:12:05.776805] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:16495 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.776823] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.785206] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f6458 00:26:12.179 [2024-07-15 18:12:05.785910] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:23151 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.785929] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.794395] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e4578 00:26:12.179 [2024-07-15 18:12:05.795096] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:23497 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.795116] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:86 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.803475] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e5658 00:26:12.179 [2024-07-15 18:12:05.804195] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:20337 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.804214] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:72 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.812563] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e6738 00:26:12.179 [2024-07-15 18:12:05.813261] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:8861 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.813280] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:25 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.821671] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190ed920 00:26:12.179 [2024-07-15 18:12:05.822381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:23238 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.822400] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:24 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.830805] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190eea00 00:26:12.179 [2024-07-15 18:12:05.831535] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:12197 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.831554] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:23 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.839875] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190efae0 00:26:12.179 [2024-07-15 18:12:05.840602] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:22686 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.840621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:22 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.848932] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f0bc0 00:26:12.179 [2024-07-15 18:12:05.849672] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:22527 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.179 [2024-07-15 18:12:05.849691] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:21 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.179 [2024-07-15 18:12:05.858214] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f1ca0 00:26:12.179 [2024-07-15 18:12:05.858935] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:17761 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.180 [2024-07-15 18:12:05.858955] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:20 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.180 [2024-07-15 18:12:05.867477] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190feb58 00:26:12.180 [2024-07-15 18:12:05.868173] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:12187 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.180 [2024-07-15 18:12:05.868192] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:19 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.180 [2024-07-15 18:12:05.876547] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190fe2e8 00:26:12.180 [2024-07-15 18:12:05.877248] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:17827 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.180 [2024-07-15 18:12:05.877267] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:18 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.180 [2024-07-15 18:12:05.885704] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f4b08 00:26:12.180 [2024-07-15 18:12:05.886408] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:3247 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.180 [2024-07-15 18:12:05.886427] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:17 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.180 [2024-07-15 18:12:05.894749] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f3a28 00:26:12.180 [2024-07-15 18:12:05.895454] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:3348 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.180 [2024-07-15 18:12:05.895473] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:16 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.180 [2024-07-15 18:12:05.903917] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190fd640 00:26:12.180 [2024-07-15 18:12:05.904647] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17953 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.180 [2024-07-15 18:12:05.904666] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.439 [2024-07-15 18:12:05.913169] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f92c0 00:26:12.439 [2024-07-15 18:12:05.913894] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:12399 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.439 [2024-07-15 18:12:05.913913] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:79 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.439 [2024-07-15 18:12:05.922278] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f81e0 00:26:12.439 [2024-07-15 18:12:05.922976] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:8441 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.439 [2024-07-15 18:12:05.922995] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:80 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.439 [2024-07-15 18:12:05.931450] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f7100 00:26:12.439 [2024-07-15 18:12:05.932159] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:11880 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.439 [2024-07-15 18:12:05.932178] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:82 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.439 [2024-07-15 18:12:05.940589] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e38d0 00:26:12.439 [2024-07-15 18:12:05.941292] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:24792 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.439 [2024-07-15 18:12:05.941311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:83 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.439 [2024-07-15 18:12:05.949701] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e49b0 00:26:12.439 [2024-07-15 18:12:05.950410] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:6049 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.439 [2024-07-15 18:12:05.950428] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:84 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.439 [2024-07-15 18:12:05.958779] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e5a90 00:26:12.439 [2024-07-15 18:12:05.959512] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:9051 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.439 [2024-07-15 18:12:05.959531] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:73 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.439 [2024-07-15 18:12:05.967992] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f5be8 00:26:12.439 [2024-07-15 18:12:05.968697] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:3477 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.439 [2024-07-15 18:12:05.968716] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:33 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.439 [2024-07-15 18:12:05.977062] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190edd58 00:26:12.439 [2024-07-15 18:12:05.977770] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:3733 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.439 [2024-07-15 18:12:05.977793] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:34 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.439 [2024-07-15 18:12:05.986237] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190eee38 00:26:12.439 [2024-07-15 18:12:05.986951] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:4233 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.439 [2024-07-15 18:12:05.986970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:35 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.439 [2024-07-15 18:12:05.995483] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190eff18 00:26:12.439 [2024-07-15 18:12:05.996203] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:22141 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.439 [2024-07-15 18:12:05.996222] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:36 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.439 [2024-07-15 18:12:06.004739] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f0ff8 00:26:12.439 [2024-07-15 18:12:06.005460] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:1153 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.439 [2024-07-15 18:12:06.005479] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:37 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.439 [2024-07-15 18:12:06.014001] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f20d8 00:26:12.439 [2024-07-15 18:12:06.014727] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:347 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.439 [2024-07-15 18:12:06.014746] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:38 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.439 [2024-07-15 18:12:06.023151] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190ff3c8 00:26:12.439 [2024-07-15 18:12:06.023889] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:5415 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.439 [2024-07-15 18:12:06.023908] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.439 [2024-07-15 18:12:06.032257] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f57b0 00:26:12.439 [2024-07-15 18:12:06.032955] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:25404 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.439 [2024-07-15 18:12:06.032974] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:40 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.439 [2024-07-15 18:12:06.041299] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f46d0 00:26:12.439 [2024-07-15 18:12:06.041995] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:8388 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.439 [2024-07-15 18:12:06.042014] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:41 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.440 [2024-07-15 18:12:06.050483] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f2948 00:26:12.440 [2024-07-15 18:12:06.051200] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:24671 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.440 [2024-07-15 18:12:06.051220] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:42 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.440 [2024-07-15 18:12:06.059628] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f96f8 00:26:12.440 [2024-07-15 18:12:06.060336] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:7723 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.440 [2024-07-15 18:12:06.060355] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:43 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.440 [2024-07-15 18:12:06.068682] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f8618 00:26:12.440 [2024-07-15 18:12:06.069389] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:21155 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.440 [2024-07-15 18:12:06.069408] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:44 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.440 [2024-07-15 18:12:06.077741] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f7538 00:26:12.440 [2024-07-15 18:12:06.078474] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:13734 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.440 [2024-07-15 18:12:06.078494] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:45 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.440 [2024-07-15 18:12:06.086905] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f6458 00:26:12.440 [2024-07-15 18:12:06.087627] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:20994 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.440 [2024-07-15 18:12:06.087646] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:26 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.440 [2024-07-15 18:12:06.095962] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e4578 00:26:12.440 [2024-07-15 18:12:06.096674] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:1266 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.440 [2024-07-15 18:12:06.096693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.440 [2024-07-15 18:12:06.105015] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e5658 00:26:12.440 [2024-07-15 18:12:06.105754] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:16367 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.440 [2024-07-15 18:12:06.105773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:101 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.440 [2024-07-15 18:12:06.114332] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e6738 00:26:12.440 [2024-07-15 18:12:06.115045] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8736 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.440 [2024-07-15 18:12:06.115063] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.440 [2024-07-15 18:12:06.123599] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190ed920 00:26:12.440 [2024-07-15 18:12:06.124329] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:13784 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.440 [2024-07-15 18:12:06.124349] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:46 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.440 [2024-07-15 18:12:06.132703] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190eea00 00:26:12.440 [2024-07-15 18:12:06.133407] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:12686 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.440 [2024-07-15 18:12:06.133426] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:85 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.440 [2024-07-15 18:12:06.141763] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190efae0 00:26:12.440 [2024-07-15 18:12:06.142473] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:23178 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.440 [2024-07-15 18:12:06.142492] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:115 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.440 [2024-07-15 18:12:06.150871] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f0bc0 00:26:12.440 [2024-07-15 18:12:06.151607] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:17110 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.440 [2024-07-15 18:12:06.151626] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:27 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.440 [2024-07-15 18:12:06.160019] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f1ca0 00:26:12.440 [2024-07-15 18:12:06.160730] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:5798 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.440 [2024-07-15 18:12:06.160749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:28 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.699 [2024-07-15 18:12:06.169284] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190feb58 00:26:12.699 [2024-07-15 18:12:06.170004] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:15548 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.699 [2024-07-15 18:12:06.170025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:29 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.699 [2024-07-15 18:12:06.178401] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190fe2e8 00:26:12.699 [2024-07-15 18:12:06.179101] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:7184 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.699 [2024-07-15 18:12:06.179119] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:30 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.699 [2024-07-15 18:12:06.187595] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f4b08 00:26:12.699 [2024-07-15 18:12:06.188298] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:19489 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.699 [2024-07-15 18:12:06.188317] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:31 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.699 [2024-07-15 18:12:06.196660] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f3a28 00:26:12.699 [2024-07-15 18:12:06.197368] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:9311 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.699 [2024-07-15 18:12:06.197389] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.699 [2024-07-15 18:12:06.205792] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190fd640 00:26:12.699 [2024-07-15 18:12:06.206522] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:4590 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.699 [2024-07-15 18:12:06.206542] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.699 [2024-07-15 18:12:06.215046] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f92c0 00:26:12.699 [2024-07-15 18:12:06.215781] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:10 nsid:1 lba:20044 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.699 [2024-07-15 18:12:06.215804] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:10 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.699 [2024-07-15 18:12:06.224305] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f81e0 00:26:12.699 [2024-07-15 18:12:06.225026] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:11 nsid:1 lba:13322 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.699 [2024-07-15 18:12:06.225046] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:11 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.699 [2024-07-15 18:12:06.233551] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f7100 00:26:12.699 [2024-07-15 18:12:06.234273] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:12 nsid:1 lba:21029 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.234293] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:12 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.242794] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e38d0 00:26:12.700 [2024-07-15 18:12:06.243519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:13 nsid:1 lba:17731 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.243539] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:13 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.252048] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e49b0 00:26:12.700 [2024-07-15 18:12:06.252774] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:2 nsid:1 lba:19771 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.252793] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:2 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.261284] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e5a90 00:26:12.700 [2024-07-15 18:12:06.262005] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:1 nsid:1 lba:17478 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.262024] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:1 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.270399] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f5be8 00:26:12.700 [2024-07-15 18:12:06.271103] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:87 nsid:1 lba:11342 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.271122] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:87 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.279449] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190edd58 00:26:12.700 [2024-07-15 18:12:06.280152] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:86 nsid:1 lba:394 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.280171] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:86 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.288609] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190eee38 00:26:12.700 [2024-07-15 18:12:06.289317] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:72 nsid:1 lba:4358 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.289336] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:72 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.297600] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190eff18 00:26:12.700 [2024-07-15 18:12:06.298394] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:25 nsid:1 lba:8255 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.298417] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:25 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.306954] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f0ff8 00:26:12.700 [2024-07-15 18:12:06.307751] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:24 nsid:1 lba:21084 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.307771] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:24 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.316219] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f20d8 00:26:12.700 [2024-07-15 18:12:06.317015] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:23 nsid:1 lba:4675 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.317034] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:23 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.325228] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190ff3c8 00:26:12.700 [2024-07-15 18:12:06.326001] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:22 nsid:1 lba:24586 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.326020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:22 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.334369] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f57b0 00:26:12.700 [2024-07-15 18:12:06.335147] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:21 nsid:1 lba:9995 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.335165] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:21 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.343584] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f46d0 00:26:12.700 [2024-07-15 18:12:06.344381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:20 nsid:1 lba:258 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.344400] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:20 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.352731] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f2948 00:26:12.700 [2024-07-15 18:12:06.353526] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:19 nsid:1 lba:9628 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.353545] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:19 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.361953] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f96f8 00:26:12.700 [2024-07-15 18:12:06.362754] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:18 nsid:1 lba:12500 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.362773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:18 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.371001] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f8618 00:26:12.700 [2024-07-15 18:12:06.371776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:17 nsid:1 lba:1394 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.371795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:17 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.380172] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f7538 00:26:12.700 [2024-07-15 18:12:06.380969] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:16 nsid:1 lba:21541 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.380988] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:16 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.389260] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f6458 00:26:12.700 [2024-07-15 18:12:06.390029] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14966 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.390046] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.398316] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e4578 00:26:12.700 [2024-07-15 18:12:06.399114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:79 nsid:1 lba:8220 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.399133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:79 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.407524] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e5658 00:26:12.700 [2024-07-15 18:12:06.408315] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:80 nsid:1 lba:2027 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.408334] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:80 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.416618] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e6738 00:26:12.700 [2024-07-15 18:12:06.417424] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:82 nsid:1 lba:3286 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.700 [2024-07-15 18:12:06.417442] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:82 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.700 [2024-07-15 18:12:06.425886] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190ed920 00:26:12.959 [2024-07-15 18:12:06.426686] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:83 nsid:1 lba:20064 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.959 [2024-07-15 18:12:06.426706] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:83 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.959 [2024-07-15 18:12:06.435103] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190eea00 00:26:12.959 [2024-07-15 18:12:06.435890] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:18129 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.959 [2024-07-15 18:12:06.435909] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:84 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.959 [2024-07-15 18:12:06.444347] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190efae0 00:26:12.959 [2024-07-15 18:12:06.445117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:73 nsid:1 lba:22686 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.959 [2024-07-15 18:12:06.445135] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:73 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.453457] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f0bc0 00:26:12.960 [2024-07-15 18:12:06.454246] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:33 nsid:1 lba:7004 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.454265] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:33 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.462541] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f1ca0 00:26:12.960 [2024-07-15 18:12:06.463313] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:34 nsid:1 lba:16281 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.463331] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:34 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.471611] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190feb58 00:26:12.960 [2024-07-15 18:12:06.472332] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:35 nsid:1 lba:18444 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.472351] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:35 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.480772] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190fe2e8 00:26:12.960 [2024-07-15 18:12:06.481479] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:36 nsid:1 lba:354 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.481498] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:36 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.489886] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f4b08 00:26:12.960 [2024-07-15 18:12:06.490668] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:37 nsid:1 lba:6439 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.490686] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:37 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.498987] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f3a28 00:26:12.960 [2024-07-15 18:12:06.499794] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:38 nsid:1 lba:5906 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.499813] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:38 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.508269] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190fd640 00:26:12.960 [2024-07-15 18:12:06.509057] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:39 nsid:1 lba:12057 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.509076] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:39 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.517527] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f92c0 00:26:12.960 [2024-07-15 18:12:06.518314] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:40 nsid:1 lba:14783 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.518333] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:40 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.526778] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f81e0 00:26:12.960 [2024-07-15 18:12:06.527574] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:41 nsid:1 lba:17118 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.527592] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:41 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.536031] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f7100 00:26:12.960 [2024-07-15 18:12:06.536841] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:42 nsid:1 lba:14953 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.536863] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:42 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.545223] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e38d0 00:26:12.960 [2024-07-15 18:12:06.546006] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:43 nsid:1 lba:8565 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.546025] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:43 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.554504] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e49b0 00:26:12.960 [2024-07-15 18:12:06.555303] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:44 nsid:1 lba:10476 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.555322] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:44 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.563670] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190e5a90 00:26:12.960 [2024-07-15 18:12:06.564395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:45 nsid:1 lba:21074 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.564414] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:45 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.572706] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f5be8 00:26:12.960 [2024-07-15 18:12:06.573477] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:26 nsid:1 lba:494 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.573495] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:26 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.581842] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190edd58 00:26:12.960 [2024-07-15 18:12:06.582622] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:69 nsid:1 lba:17610 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.582640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:69 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.590943] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190eee38 00:26:12.960 [2024-07-15 18:12:06.591736] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:101 nsid:1 lba:22481 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.591755] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:101 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.600038] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190eff18 00:26:12.960 [2024-07-15 18:12:06.600823] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:0 nsid:1 lba:8326 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.600841] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:0 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.609170] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f0ff8 00:26:12.960 [2024-07-15 18:12:06.609969] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:46 nsid:1 lba:25249 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.609989] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:46 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.618304] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f20d8 00:26:12.960 [2024-07-15 18:12:06.619078] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:85 nsid:1 lba:16628 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.619096] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:85 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.627382] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190ff3c8 00:26:12.960 [2024-07-15 18:12:06.628170] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:115 nsid:1 lba:13062 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.628188] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:115 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.636496] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f57b0 00:26:12.960 [2024-07-15 18:12:06.637267] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:27 nsid:1 lba:11184 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.637285] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:27 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.645587] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f46d0 00:26:12.960 [2024-07-15 18:12:06.646372] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:28 nsid:1 lba:7217 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.646391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:28 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.654699] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f2948 00:26:12.960 [2024-07-15 18:12:06.655477] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:29 nsid:1 lba:22015 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.655495] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:29 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.663765] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f96f8 00:26:12.960 [2024-07-15 18:12:06.664554] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:30 nsid:1 lba:14545 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.664573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:30 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.672944] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f8618 00:26:12.960 [2024-07-15 18:12:06.673757] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:31 nsid:1 lba:5426 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.673776] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:31 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:12.960 [2024-07-15 18:12:06.682177] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f7538 00:26:12.960 [2024-07-15 18:12:06.682985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:8 nsid:1 lba:1789 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:12.960 [2024-07-15 18:12:06.683004] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:8 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:13.219 [2024-07-15 18:12:06.691496] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2541710) with pdu=0x2000190f6458 00:26:13.219 [2024-07-15 18:12:06.692297] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:9 nsid:1 lba:4645 len:1 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:13.219 [2024-07-15 18:12:06.692316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:9 cdw0:0 sqhd:004f p:0 m:0 dnr:0 00:26:13.219 00:26:13.219 Latency(us) 00:26:13.219 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:13.219 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 128, IO size: 4096) 00:26:13.219 nvme0n1 : 2.00 27800.00 108.59 0.00 0.00 4597.90 1880.60 15614.66 00:26:13.219 =================================================================================================================== 00:26:13.219 Total : 27800.00 108.59 0.00 0.00 4597.90 1880.60 15614.66 00:26:13.219 0 00:26:13.219 18:12:06 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:26:13.219 18:12:06 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:26:13.219 18:12:06 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:26:13.219 | .driver_specific 00:26:13.219 | .nvme_error 00:26:13.219 | .status_code 00:26:13.219 | .command_transient_transport_error' 00:26:13.219 18:12:06 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:26:13.219 18:12:06 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # (( 218 > 0 )) 00:26:13.219 18:12:06 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@73 -- # killprocess 734510 00:26:13.219 18:12:06 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@948 -- # '[' -z 734510 ']' 00:26:13.219 18:12:06 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@952 -- # kill -0 734510 00:26:13.219 18:12:06 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@953 -- # uname 00:26:13.219 18:12:06 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:26:13.219 18:12:06 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 734510 00:26:13.219 18:12:06 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:26:13.219 18:12:06 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:26:13.219 18:12:06 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@966 -- # echo 'killing process with pid 734510' 00:26:13.219 killing process with pid 734510 00:26:13.219 18:12:06 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@967 -- # kill 734510 00:26:13.219 Received shutdown signal, test time was about 2.000000 seconds 00:26:13.219 00:26:13.219 Latency(us) 00:26:13.219 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:13.219 =================================================================================================================== 00:26:13.219 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:26:13.219 18:12:06 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@972 -- # wait 734510 00:26:13.478 18:12:07 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@115 -- # run_bperf_err randwrite 131072 16 00:26:13.478 18:12:07 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@54 -- # local rw bs qd 00:26:13.478 18:12:07 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # rw=randwrite 00:26:13.478 18:12:07 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # bs=131072 00:26:13.478 18:12:07 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@56 -- # qd=16 00:26:13.478 18:12:07 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@58 -- # bperfpid=735101 00:26:13.478 18:12:07 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@60 -- # waitforlisten 735101 /var/tmp/bperf.sock 00:26:13.478 18:12:07 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@57 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -m 2 -r /var/tmp/bperf.sock -w randwrite -o 131072 -t 2 -q 16 -z 00:26:13.478 18:12:07 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@829 -- # '[' -z 735101 ']' 00:26:13.478 18:12:07 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bperf.sock 00:26:13.478 18:12:07 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@834 -- # local max_retries=100 00:26:13.478 18:12:07 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:26:13.478 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:26:13.478 18:12:07 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@838 -- # xtrace_disable 00:26:13.478 18:12:07 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:26:13.478 [2024-07-15 18:12:07.164529] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:26:13.478 [2024-07-15 18:12:07.164576] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid735101 ] 00:26:13.478 I/O size of 131072 is greater than zero copy threshold (65536). 00:26:13.478 Zero copy mechanism will not be used. 00:26:13.478 EAL: No free 2048 kB hugepages reported on node 1 00:26:13.737 [2024-07-15 18:12:07.217536] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:13.737 [2024-07-15 18:12:07.287683] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:26:14.304 18:12:07 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:26:14.304 18:12:07 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@862 -- # return 0 00:26:14.304 18:12:07 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@61 -- # bperf_rpc bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:26:14.304 18:12:07 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_set_options --nvme-error-stat --bdev-retry-count -1 00:26:14.563 18:12:08 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@63 -- # rpc_cmd accel_error_inject_error -o crc32c -t disable 00:26:14.563 18:12:08 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:14.563 18:12:08 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:26:14.563 18:12:08 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:14.563 18:12:08 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@64 -- # bperf_rpc bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:26:14.563 18:12:08 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller --ddgst -t tcp -a 10.0.0.2 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode1 -b nvme0 00:26:14.822 nvme0n1 00:26:14.822 18:12:08 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@67 -- # rpc_cmd accel_error_inject_error -o crc32c -t corrupt -i 32 00:26:14.822 18:12:08 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:14.822 18:12:08 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:26:14.822 18:12:08 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:14.822 18:12:08 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@69 -- # bperf_py perform_tests 00:26:14.822 18:12:08 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@19 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:26:14.822 I/O size of 131072 is greater than zero copy threshold (65536). 00:26:14.822 Zero copy mechanism will not be used. 00:26:14.822 Running I/O for 2 seconds... 00:26:14.822 [2024-07-15 18:12:08.492870] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:14.822 [2024-07-15 18:12:08.492974] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:704 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:14.822 [2024-07-15 18:12:08.493008] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:14.822 [2024-07-15 18:12:08.502053] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:14.822 [2024-07-15 18:12:08.502452] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:14.822 [2024-07-15 18:12:08.502477] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:14.822 [2024-07-15 18:12:08.509148] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:14.822 [2024-07-15 18:12:08.509527] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:14.822 [2024-07-15 18:12:08.509549] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:14.822 [2024-07-15 18:12:08.515493] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:14.822 [2024-07-15 18:12:08.515866] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:14.822 [2024-07-15 18:12:08.515888] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:14.822 [2024-07-15 18:12:08.521398] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:14.822 [2024-07-15 18:12:08.521763] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:14.822 [2024-07-15 18:12:08.521784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:14.822 [2024-07-15 18:12:08.527190] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:14.822 [2024-07-15 18:12:08.527578] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14848 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:14.822 [2024-07-15 18:12:08.527600] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:14.822 [2024-07-15 18:12:08.533756] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:14.822 [2024-07-15 18:12:08.534131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:14.822 [2024-07-15 18:12:08.534152] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:14.822 [2024-07-15 18:12:08.538569] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:14.822 [2024-07-15 18:12:08.538957] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:14.822 [2024-07-15 18:12:08.538977] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:14.822 [2024-07-15 18:12:08.543233] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:14.822 [2024-07-15 18:12:08.543612] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:14.822 [2024-07-15 18:12:08.543633] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:14.822 [2024-07-15 18:12:08.548062] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:14.822 [2024-07-15 18:12:08.548450] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:14.823 [2024-07-15 18:12:08.548471] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.082 [2024-07-15 18:12:08.552914] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.082 [2024-07-15 18:12:08.553303] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.082 [2024-07-15 18:12:08.553324] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.082 [2024-07-15 18:12:08.557774] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.082 [2024-07-15 18:12:08.558148] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.082 [2024-07-15 18:12:08.558168] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.082 [2024-07-15 18:12:08.562585] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.082 [2024-07-15 18:12:08.562950] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.082 [2024-07-15 18:12:08.562970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.082 [2024-07-15 18:12:08.567369] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.082 [2024-07-15 18:12:08.567748] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.082 [2024-07-15 18:12:08.567768] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.082 [2024-07-15 18:12:08.572172] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.082 [2024-07-15 18:12:08.572554] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.082 [2024-07-15 18:12:08.572573] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.082 [2024-07-15 18:12:08.576803] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.082 [2024-07-15 18:12:08.577171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.082 [2024-07-15 18:12:08.577192] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.082 [2024-07-15 18:12:08.581413] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.082 [2024-07-15 18:12:08.581800] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.082 [2024-07-15 18:12:08.581819] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.082 [2024-07-15 18:12:08.586108] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.082 [2024-07-15 18:12:08.586503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.082 [2024-07-15 18:12:08.586523] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.082 [2024-07-15 18:12:08.590770] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.082 [2024-07-15 18:12:08.591163] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.082 [2024-07-15 18:12:08.591183] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.082 [2024-07-15 18:12:08.595610] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.082 [2024-07-15 18:12:08.595983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.082 [2024-07-15 18:12:08.596003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.082 [2024-07-15 18:12:08.601004] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.082 [2024-07-15 18:12:08.601401] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10976 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.082 [2024-07-15 18:12:08.601422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.082 [2024-07-15 18:12:08.605718] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.082 [2024-07-15 18:12:08.606101] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18368 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.082 [2024-07-15 18:12:08.606122] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.082 [2024-07-15 18:12:08.610535] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.082 [2024-07-15 18:12:08.610915] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.082 [2024-07-15 18:12:08.610936] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.082 [2024-07-15 18:12:08.615908] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.082 [2024-07-15 18:12:08.616290] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.082 [2024-07-15 18:12:08.616310] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.082 [2024-07-15 18:12:08.620722] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.082 [2024-07-15 18:12:08.621118] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.082 [2024-07-15 18:12:08.621138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.082 [2024-07-15 18:12:08.625595] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.082 [2024-07-15 18:12:08.625981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.082 [2024-07-15 18:12:08.626001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.082 [2024-07-15 18:12:08.630365] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.082 [2024-07-15 18:12:08.630739] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.082 [2024-07-15 18:12:08.630761] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.082 [2024-07-15 18:12:08.635091] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.082 [2024-07-15 18:12:08.635477] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.082 [2024-07-15 18:12:08.635497] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.082 [2024-07-15 18:12:08.639683] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.082 [2024-07-15 18:12:08.640074] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.082 [2024-07-15 18:12:08.640094] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.082 [2024-07-15 18:12:08.645344] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.082 [2024-07-15 18:12:08.645734] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.082 [2024-07-15 18:12:08.645754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.083 [2024-07-15 18:12:08.650116] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.083 [2024-07-15 18:12:08.650489] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.083 [2024-07-15 18:12:08.650509] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.083 [2024-07-15 18:12:08.654923] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.083 [2024-07-15 18:12:08.655309] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.083 [2024-07-15 18:12:08.655328] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.083 [2024-07-15 18:12:08.660282] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.083 [2024-07-15 18:12:08.660666] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.083 [2024-07-15 18:12:08.660685] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.083 [2024-07-15 18:12:08.665980] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.083 [2024-07-15 18:12:08.666389] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.083 [2024-07-15 18:12:08.666410] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.083 [2024-07-15 18:12:08.671348] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.083 [2024-07-15 18:12:08.671726] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.083 [2024-07-15 18:12:08.671746] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.083 [2024-07-15 18:12:08.678552] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.083 [2024-07-15 18:12:08.678941] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.083 [2024-07-15 18:12:08.678961] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.083 [2024-07-15 18:12:08.686096] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.083 [2024-07-15 18:12:08.686399] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.083 [2024-07-15 18:12:08.686419] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.083 [2024-07-15 18:12:08.694626] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.083 [2024-07-15 18:12:08.695022] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.083 [2024-07-15 18:12:08.695041] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.083 [2024-07-15 18:12:08.703236] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.083 [2024-07-15 18:12:08.703622] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10880 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.083 [2024-07-15 18:12:08.703642] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.083 [2024-07-15 18:12:08.711926] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.083 [2024-07-15 18:12:08.712095] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.083 [2024-07-15 18:12:08.712113] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.083 [2024-07-15 18:12:08.721521] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.083 [2024-07-15 18:12:08.721925] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.083 [2024-07-15 18:12:08.721945] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.083 [2024-07-15 18:12:08.731820] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.083 [2024-07-15 18:12:08.732210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.083 [2024-07-15 18:12:08.732235] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.083 [2024-07-15 18:12:08.740981] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.083 [2024-07-15 18:12:08.741395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.083 [2024-07-15 18:12:08.741414] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.083 [2024-07-15 18:12:08.750399] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.083 [2024-07-15 18:12:08.750799] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.083 [2024-07-15 18:12:08.750819] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.083 [2024-07-15 18:12:08.759251] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.083 [2024-07-15 18:12:08.759658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5760 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.083 [2024-07-15 18:12:08.759677] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.083 [2024-07-15 18:12:08.768700] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.083 [2024-07-15 18:12:08.769114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.083 [2024-07-15 18:12:08.769133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.083 [2024-07-15 18:12:08.777423] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.083 [2024-07-15 18:12:08.777826] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.083 [2024-07-15 18:12:08.777846] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.083 [2024-07-15 18:12:08.786036] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.083 [2024-07-15 18:12:08.786462] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.083 [2024-07-15 18:12:08.786483] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.083 [2024-07-15 18:12:08.796343] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.083 [2024-07-15 18:12:08.796754] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.083 [2024-07-15 18:12:08.796773] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.083 [2024-07-15 18:12:08.804818] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.083 [2024-07-15 18:12:08.805212] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3424 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.083 [2024-07-15 18:12:08.805239] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.342 [2024-07-15 18:12:08.812145] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.342 [2024-07-15 18:12:08.812544] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22080 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.342 [2024-07-15 18:12:08.812565] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.342 [2024-07-15 18:12:08.819297] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.342 [2024-07-15 18:12:08.819369] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.342 [2024-07-15 18:12:08.819388] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:08.827000] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:08.827398] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:08.827422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:08.834438] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:08.834829] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:08.834849] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:08.842563] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:08.842713] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:08.842731] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:08.851941] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:08.852345] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:08.852365] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:08.859888] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:08.860312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:08.860332] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:08.868686] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:08.869077] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:08.869097] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:08.877647] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:08.878043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:08.878062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:08.886959] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:08.887377] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:08.887397] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:08.896593] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:08.897000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:08.897020] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:08.906269] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:08.906693] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:08.906712] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:08.915593] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:08.916000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15168 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:08.916019] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:08.925381] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:08.925765] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:08.925784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:08.934311] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:08.934706] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:08.934725] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:08.942480] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:08.942885] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:08.942905] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:08.951649] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:08.952044] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16000 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:08.952063] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:08.960511] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:08.960756] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:08.960776] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:08.969135] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:08.969551] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:08.969571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:08.977593] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:08.977992] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:08.978012] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:08.986053] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:08.986438] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19264 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:08.986458] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:08.995354] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:08.995781] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:08.995801] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:09.004068] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:09.004503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19200 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:09.004523] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:09.011886] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:09.012296] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:64 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:09.012316] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:09.018525] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:09.018906] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:64 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:09.018926] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:09.026044] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:09.026453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22176 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:09.026473] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:09.033374] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:09.033775] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:09.033795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:09.041969] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:09.042391] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:09.042411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:09.049881] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:09.050262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:09.050285] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:09.058841] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:09.059255] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:09.059274] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.343 [2024-07-15 18:12:09.068222] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.343 [2024-07-15 18:12:09.068632] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.343 [2024-07-15 18:12:09.068651] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.076504] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.076729] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.076749] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.084992] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.085529] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.085548] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.093955] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.094403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.094434] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.102588] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.103071] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.103090] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.110639] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.111144] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.111164] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.117108] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.117455] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.117475] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.122414] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.122780] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.122799] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.127410] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.127776] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.127795] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.132879] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.133231] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11712 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.133251] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.137493] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.137859] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.137878] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.142190] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.142565] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.142584] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.146752] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.147097] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7072 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.147116] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.151507] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.151875] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.151894] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.156561] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.156923] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.156941] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.161321] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.161672] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14880 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.161694] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.166620] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.166985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.167004] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.172340] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.172691] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.172710] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.178454] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.178818] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.178838] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.184240] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.184616] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.184635] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.190777] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.191135] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.191154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.197046] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.197403] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.197422] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.202835] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.203194] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.203214] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.207843] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.208193] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.208213] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.213051] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.213440] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.213459] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.217821] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.218182] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.218201] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.222674] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.223063] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.223082] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.227703] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.228068] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.228088] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.233267] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.604 [2024-07-15 18:12:09.233636] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.604 [2024-07-15 18:12:09.233655] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.604 [2024-07-15 18:12:09.237895] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.605 [2024-07-15 18:12:09.238272] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.605 [2024-07-15 18:12:09.238291] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.605 [2024-07-15 18:12:09.242742] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.605 [2024-07-15 18:12:09.243124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.605 [2024-07-15 18:12:09.243143] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.605 [2024-07-15 18:12:09.247470] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.605 [2024-07-15 18:12:09.247834] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7104 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.605 [2024-07-15 18:12:09.247854] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.605 [2024-07-15 18:12:09.252081] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.605 [2024-07-15 18:12:09.252446] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.605 [2024-07-15 18:12:09.252468] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.605 [2024-07-15 18:12:09.256695] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.605 [2024-07-15 18:12:09.257059] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23040 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.605 [2024-07-15 18:12:09.257080] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.605 [2024-07-15 18:12:09.261326] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.605 [2024-07-15 18:12:09.261690] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17664 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.605 [2024-07-15 18:12:09.261710] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.605 [2024-07-15 18:12:09.266021] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.605 [2024-07-15 18:12:09.266376] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.605 [2024-07-15 18:12:09.266396] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.605 [2024-07-15 18:12:09.271260] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.605 [2024-07-15 18:12:09.271632] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.605 [2024-07-15 18:12:09.271651] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.605 [2024-07-15 18:12:09.276535] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.605 [2024-07-15 18:12:09.276897] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.605 [2024-07-15 18:12:09.276916] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.605 [2024-07-15 18:12:09.282020] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.605 [2024-07-15 18:12:09.282391] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.605 [2024-07-15 18:12:09.282411] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.605 [2024-07-15 18:12:09.288195] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.605 [2024-07-15 18:12:09.288569] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16832 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.605 [2024-07-15 18:12:09.288589] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.605 [2024-07-15 18:12:09.294367] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.605 [2024-07-15 18:12:09.294728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.605 [2024-07-15 18:12:09.294747] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.605 [2024-07-15 18:12:09.300127] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.605 [2024-07-15 18:12:09.300498] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.605 [2024-07-15 18:12:09.300521] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.605 [2024-07-15 18:12:09.306331] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.605 [2024-07-15 18:12:09.306705] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.605 [2024-07-15 18:12:09.306725] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.605 [2024-07-15 18:12:09.313106] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.605 [2024-07-15 18:12:09.313473] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.605 [2024-07-15 18:12:09.313493] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.605 [2024-07-15 18:12:09.319291] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.605 [2024-07-15 18:12:09.319663] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.605 [2024-07-15 18:12:09.319682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.605 [2024-07-15 18:12:09.325655] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.605 [2024-07-15 18:12:09.326003] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.605 [2024-07-15 18:12:09.326024] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.864 [2024-07-15 18:12:09.332907] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.864 [2024-07-15 18:12:09.333419] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.864 [2024-07-15 18:12:09.333439] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.864 [2024-07-15 18:12:09.343831] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.864 [2024-07-15 18:12:09.344337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.864 [2024-07-15 18:12:09.344358] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.864 [2024-07-15 18:12:09.351664] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.864 [2024-07-15 18:12:09.352113] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.864 [2024-07-15 18:12:09.352132] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.864 [2024-07-15 18:12:09.358760] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.359124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.359143] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.364486] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.364855] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.364874] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.371045] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.371519] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25408 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.371538] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.379040] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.379452] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.379471] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.386503] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.387038] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.387073] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.394899] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.395355] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.395374] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.403117] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.403559] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.403579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.411307] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.411650] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24416 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.411668] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.418669] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.419132] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.419150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.426219] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.426738] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.426758] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.434447] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.434888] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.434907] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.442811] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.443195] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.443216] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.450995] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.451455] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.451474] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.459238] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.459738] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.459758] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.467621] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.468100] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.468120] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.475785] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.476244] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.476263] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.484120] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.484575] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.484595] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.492110] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.492582] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17952 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.492601] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.500564] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.500962] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.500983] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.508293] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.508743] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20480 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.508763] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.515845] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.516270] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.516289] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.523626] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.523957] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.523977] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.531246] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.531642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.531661] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.538293] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.538716] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.538734] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.545692] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.546112] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.546131] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.553490] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.553926] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.553944] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.561338] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.865 [2024-07-15 18:12:09.561766] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8768 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.865 [2024-07-15 18:12:09.561784] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:15.865 [2024-07-15 18:12:09.568436] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.866 [2024-07-15 18:12:09.568890] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.866 [2024-07-15 18:12:09.568909] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:15.866 [2024-07-15 18:12:09.576116] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.866 [2024-07-15 18:12:09.576549] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.866 [2024-07-15 18:12:09.576568] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:15.866 [2024-07-15 18:12:09.584309] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:15.866 [2024-07-15 18:12:09.584732] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:15.866 [2024-07-15 18:12:09.584751] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.125 [2024-07-15 18:12:09.592166] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.125 [2024-07-15 18:12:09.592591] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23328 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.125 [2024-07-15 18:12:09.592610] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.125 [2024-07-15 18:12:09.600179] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.125 [2024-07-15 18:12:09.600588] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21728 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.125 [2024-07-15 18:12:09.600606] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.125 [2024-07-15 18:12:09.607939] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.125 [2024-07-15 18:12:09.608387] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:576 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.125 [2024-07-15 18:12:09.608407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.125 [2024-07-15 18:12:09.615503] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.125 [2024-07-15 18:12:09.615898] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.125 [2024-07-15 18:12:09.615917] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.125 [2024-07-15 18:12:09.623300] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.125 [2024-07-15 18:12:09.623718] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.125 [2024-07-15 18:12:09.623737] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.125 [2024-07-15 18:12:09.630724] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.631138] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15584 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.631160] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.638162] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.638602] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.638621] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.646780] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.647187] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.647207] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.654339] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.654728] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.654747] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.661746] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.662115] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15456 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.662133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.668656] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.669059] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22080 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.669078] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.676174] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.676613] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20928 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.676632] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.683971] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.684393] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.684412] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.691940] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.692312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.692331] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.698209] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.698553] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15840 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.698572] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.704221] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.704548] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.704566] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.710611] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.710942] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.710961] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.716638] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.716988] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.717007] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.722258] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.722621] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.722640] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.728083] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.728439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.728458] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.734515] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.734842] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.734861] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.740169] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.740488] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18464 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.740507] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.745797] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.746150] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14912 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.746169] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.751710] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.752022] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.752041] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.757197] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.757525] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.757545] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.762955] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.763274] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.763295] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.769013] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.769336] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10080 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.769357] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.774702] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.775043] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.775062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.780475] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.780806] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9248 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.780825] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.786206] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.786542] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.786561] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.791731] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.792060] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.792079] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.797627] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.797947] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.797970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.803352] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.803700] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.803720] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.809494] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.126 [2024-07-15 18:12:09.809814] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2400 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.126 [2024-07-15 18:12:09.809833] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.126 [2024-07-15 18:12:09.814934] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.127 [2024-07-15 18:12:09.815286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.127 [2024-07-15 18:12:09.815305] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.127 [2024-07-15 18:12:09.820636] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.127 [2024-07-15 18:12:09.820957] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.127 [2024-07-15 18:12:09.820975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.127 [2024-07-15 18:12:09.826009] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.127 [2024-07-15 18:12:09.826336] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12096 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.127 [2024-07-15 18:12:09.826355] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.127 [2024-07-15 18:12:09.830488] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.127 [2024-07-15 18:12:09.830814] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4544 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.127 [2024-07-15 18:12:09.830833] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.127 [2024-07-15 18:12:09.834986] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.127 [2024-07-15 18:12:09.835286] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11904 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.127 [2024-07-15 18:12:09.835305] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.127 [2024-07-15 18:12:09.839349] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.127 [2024-07-15 18:12:09.839632] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.127 [2024-07-15 18:12:09.839651] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.127 [2024-07-15 18:12:09.843407] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.127 [2024-07-15 18:12:09.843684] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.127 [2024-07-15 18:12:09.843703] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.127 [2024-07-15 18:12:09.847359] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.127 [2024-07-15 18:12:09.847642] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.127 [2024-07-15 18:12:09.847661] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.127 [2024-07-15 18:12:09.851280] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.127 [2024-07-15 18:12:09.851560] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25472 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.127 [2024-07-15 18:12:09.851579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.387 [2024-07-15 18:12:09.855180] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.387 [2024-07-15 18:12:09.855444] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19680 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.387 [2024-07-15 18:12:09.855463] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.387 [2024-07-15 18:12:09.859112] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.387 [2024-07-15 18:12:09.859412] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19968 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.387 [2024-07-15 18:12:09.859431] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.387 [2024-07-15 18:12:09.863361] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.387 [2024-07-15 18:12:09.863639] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.387 [2024-07-15 18:12:09.863657] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.387 [2024-07-15 18:12:09.868468] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.387 [2024-07-15 18:12:09.868786] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15744 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.387 [2024-07-15 18:12:09.868806] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.387 [2024-07-15 18:12:09.873156] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.387 [2024-07-15 18:12:09.873439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.387 [2024-07-15 18:12:09.873458] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.387 [2024-07-15 18:12:09.877722] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.387 [2024-07-15 18:12:09.877993] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.387 [2024-07-15 18:12:09.878013] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.387 [2024-07-15 18:12:09.882070] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.387 [2024-07-15 18:12:09.882337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10944 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.387 [2024-07-15 18:12:09.882355] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.387 [2024-07-15 18:12:09.886686] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.387 [2024-07-15 18:12:09.886963] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8576 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.387 [2024-07-15 18:12:09.886982] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.387 [2024-07-15 18:12:09.890922] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.387 [2024-07-15 18:12:09.891155] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:32 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.387 [2024-07-15 18:12:09.891174] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.387 [2024-07-15 18:12:09.895557] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.387 [2024-07-15 18:12:09.895810] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20992 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.387 [2024-07-15 18:12:09.895829] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.899743] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.899996] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24064 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.900015] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.903860] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.904119] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.904138] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.908105] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.908366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.908385] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.912978] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.913250] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.913268] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.918253] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.918503] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.918526] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.922666] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.922931] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19168 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.922951] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.927303] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.927560] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.927579] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.931992] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.932278] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.932297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.936465] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.936663] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.936682] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.940200] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.940406] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.940426] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.943986] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.944237] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.944256] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.947717] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.947914] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17312 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.947932] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.951434] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.951636] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16288 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.951655] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.955108] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.955308] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:11520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.955326] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.959306] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.959516] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.959535] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.963545] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.963767] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.963786] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.967783] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.967981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20640 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.968000] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.971904] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.972114] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.972133] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.976221] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.976431] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.976457] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.980668] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.980870] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1600 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.980889] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.984818] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.985021] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13376 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.985048] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.989345] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.989543] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.989571] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.993504] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.993706] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9152 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.993724] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:09.997785] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:09.997985] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:09.998003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:10.002124] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:10.002325] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:10.002345] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:10.006458] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:10.006665] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:10.006683] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:10.010504] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:10.010706] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:7872 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:10.010724] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:10.015018] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:10.015218] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:10.015241] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.388 [2024-07-15 18:12:10.019202] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.388 [2024-07-15 18:12:10.019419] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.388 [2024-07-15 18:12:10.019449] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.389 [2024-07-15 18:12:10.023366] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.389 [2024-07-15 18:12:10.023570] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18880 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.389 [2024-07-15 18:12:10.023589] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.389 [2024-07-15 18:12:10.027275] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.389 [2024-07-15 18:12:10.027484] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.389 [2024-07-15 18:12:10.027503] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.389 [2024-07-15 18:12:10.032916] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.389 [2024-07-15 18:12:10.033131] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:14240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.389 [2024-07-15 18:12:10.033154] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.389 [2024-07-15 18:12:10.036889] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.389 [2024-07-15 18:12:10.037093] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.389 [2024-07-15 18:12:10.037112] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.389 [2024-07-15 18:12:10.041146] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.389 [2024-07-15 18:12:10.041356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:1184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.389 [2024-07-15 18:12:10.041376] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.389 [2024-07-15 18:12:10.046665] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.389 [2024-07-15 18:12:10.046872] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:8960 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.389 [2024-07-15 18:12:10.046891] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.389 [2024-07-15 18:12:10.051906] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.389 [2024-07-15 18:12:10.052180] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.389 [2024-07-15 18:12:10.052200] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.389 [2024-07-15 18:12:10.058911] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.389 [2024-07-15 18:12:10.059139] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.389 [2024-07-15 18:12:10.059158] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.389 [2024-07-15 18:12:10.064989] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.389 [2024-07-15 18:12:10.065262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23392 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.389 [2024-07-15 18:12:10.065282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.389 [2024-07-15 18:12:10.071814] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.389 [2024-07-15 18:12:10.072191] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18304 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.389 [2024-07-15 18:12:10.072232] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.389 [2024-07-15 18:12:10.078763] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.389 [2024-07-15 18:12:10.079050] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.389 [2024-07-15 18:12:10.079070] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.389 [2024-07-15 18:12:10.085544] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.389 [2024-07-15 18:12:10.085838] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.389 [2024-07-15 18:12:10.085858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.389 [2024-07-15 18:12:10.092090] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.389 [2024-07-15 18:12:10.092388] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.389 [2024-07-15 18:12:10.092407] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.389 [2024-07-15 18:12:10.098887] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.389 [2024-07-15 18:12:10.099207] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9024 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.389 [2024-07-15 18:12:10.099234] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.389 [2024-07-15 18:12:10.105977] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.389 [2024-07-15 18:12:10.106240] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20032 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.389 [2024-07-15 18:12:10.106260] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.389 [2024-07-15 18:12:10.112875] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.389 [2024-07-15 18:12:10.113123] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.389 [2024-07-15 18:12:10.113143] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.649 [2024-07-15 18:12:10.120577] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.649 [2024-07-15 18:12:10.120881] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25536 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.649 [2024-07-15 18:12:10.120901] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.649 [2024-07-15 18:12:10.128109] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.649 [2024-07-15 18:12:10.128365] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.649 [2024-07-15 18:12:10.128384] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.649 [2024-07-15 18:12:10.136131] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.649 [2024-07-15 18:12:10.136356] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.649 [2024-07-15 18:12:10.136391] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.649 [2024-07-15 18:12:10.144070] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.649 [2024-07-15 18:12:10.144393] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20800 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.649 [2024-07-15 18:12:10.144414] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.649 [2024-07-15 18:12:10.151989] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.649 [2024-07-15 18:12:10.152316] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3520 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.649 [2024-07-15 18:12:10.152336] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.649 [2024-07-15 18:12:10.159737] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.649 [2024-07-15 18:12:10.159982] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4352 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.649 [2024-07-15 18:12:10.160003] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.649 [2024-07-15 18:12:10.168156] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.649 [2024-07-15 18:12:10.168440] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.649 [2024-07-15 18:12:10.168470] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.649 [2024-07-15 18:12:10.175987] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.649 [2024-07-15 18:12:10.176257] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22240 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.649 [2024-07-15 18:12:10.176278] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.649 [2024-07-15 18:12:10.183916] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.184197] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5216 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.184218] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.191875] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.192093] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.192114] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.199711] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.200042] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3296 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.200063] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.207876] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.208134] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19008 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.208155] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.215850] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.216108] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.216128] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.223078] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.223347] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.223367] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.229112] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.229405] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.229426] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.234762] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.235047] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:15136 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.235068] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.239283] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.239539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3808 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.239559] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.244000] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.244280] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6560 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.244301] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.249645] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.249951] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5184 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.249971] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.254454] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.254697] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24384 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.254717] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.258848] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.259087] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5888 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.259108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.263410] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.263658] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17120 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.263679] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.267783] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.267977] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18592 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.267999] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.271966] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.272221] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.272250] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.277072] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.277312] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13504 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.277333] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.282391] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.282673] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.282693] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.288394] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.288714] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24160 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.288734] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.294098] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.294368] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3552 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.294389] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.301215] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.301542] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25344 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.301565] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.307335] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.307566] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.307587] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.312249] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.312475] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22432 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.312495] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.316531] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.316758] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:13792 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.316778] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.320821] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.321026] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10048 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.321047] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.324999] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.325210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12128 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.325236] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.329125] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.329314] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:3232 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.329333] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.333466] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.333655] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6112 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.333675] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.650 [2024-07-15 18:12:10.338914] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.650 [2024-07-15 18:12:10.339117] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:22656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.650 [2024-07-15 18:12:10.339136] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.651 [2024-07-15 18:12:10.343568] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.651 [2024-07-15 18:12:10.343760] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20672 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.651 [2024-07-15 18:12:10.343781] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.651 [2024-07-15 18:12:10.348106] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.651 [2024-07-15 18:12:10.348300] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6016 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.651 [2024-07-15 18:12:10.348321] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.651 [2024-07-15 18:12:10.352590] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.651 [2024-07-15 18:12:10.352779] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25440 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.651 [2024-07-15 18:12:10.352800] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.651 [2024-07-15 18:12:10.356973] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.651 [2024-07-15 18:12:10.357171] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.651 [2024-07-15 18:12:10.357190] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.651 [2024-07-15 18:12:10.361282] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.651 [2024-07-15 18:12:10.361487] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:20736 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.651 [2024-07-15 18:12:10.361508] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.651 [2024-07-15 18:12:10.366114] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.651 [2024-07-15 18:12:10.366317] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.651 [2024-07-15 18:12:10.366338] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.651 [2024-07-15 18:12:10.372000] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.651 [2024-07-15 18:12:10.372208] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12864 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.651 [2024-07-15 18:12:10.372236] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.911 [2024-07-15 18:12:10.376409] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.911 [2024-07-15 18:12:10.376614] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:12608 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.911 [2024-07-15 18:12:10.376635] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.911 [2024-07-15 18:12:10.380796] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.911 [2024-07-15 18:12:10.380981] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:4224 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.911 [2024-07-15 18:12:10.381001] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.911 [2024-07-15 18:12:10.385162] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.911 [2024-07-15 18:12:10.385373] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16896 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.911 [2024-07-15 18:12:10.385393] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.389530] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.389718] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19488 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.389739] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.393810] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.394020] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23776 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.394040] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.398155] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.398366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17984 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.398385] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.402562] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.402762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5056 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.402783] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.406899] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.407088] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:25280 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.407108] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.411166] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.411372] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9824 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.411392] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.415816] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.415999] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6816 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.416018] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.420472] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.420717] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:16448 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.420745] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.424967] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.425210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:10720 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.425237] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.430432] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.430668] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:24256 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.430687] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.435380] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.435626] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:19616 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.435645] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.439715] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.439950] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.439970] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.443992] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.444233] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:2784 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.444251] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.448341] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.448592] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:23936 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.448611] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.452444] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.452690] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:17632 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.452708] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.456360] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.456604] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:5088 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.456623] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.460263] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.460509] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:9696 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.460528] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.464143] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.464383] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18336 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.464401] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.468026] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.468281] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6656 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.468300] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0061 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.472309] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.472547] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:18848 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.472566] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0001 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.476940] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.477167] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:21568 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.477186] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0021 p:0 m:0 dnr:0 00:26:16.912 [2024-07-15 18:12:10.482125] tcp.c:2123:data_crc32_calc_done: *ERROR*: Data digest error on tqpair=(0x2376890) with pdu=0x2000190fef90 00:26:16.912 [2024-07-15 18:12:10.482189] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:15 nsid:1 lba:6208 len:32 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:16.912 [2024-07-15 18:12:10.482207] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: COMMAND TRANSIENT TRANSPORT ERROR (00/22) qid:1 cid:15 cdw0:0 sqhd:0041 p:0 m:0 dnr:0 00:26:16.912 00:26:16.912 Latency(us) 00:26:16.912 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:16.912 Job: nvme0n1 (Core Mask 0x2, workload: randwrite, depth: 16, IO size: 131072) 00:26:16.912 nvme0n1 : 2.00 5049.15 631.14 0.00 0.00 3164.22 1752.38 13221.18 00:26:16.912 =================================================================================================================== 00:26:16.912 Total : 5049.15 631.14 0.00 0.00 3164.22 1752.38 13221.18 00:26:16.912 0 00:26:16.912 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # get_transient_errcount nvme0n1 00:26:16.912 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@27 -- # bperf_rpc bdev_get_iostat -b nvme0n1 00:26:16.912 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@28 -- # jq -r '.bdevs[0] 00:26:16.912 | .driver_specific 00:26:16.912 | .nvme_error 00:26:16.912 | .status_code 00:26:16.912 | .command_transient_transport_error' 00:26:16.912 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@18 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_get_iostat -b nvme0n1 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@71 -- # (( 326 > 0 )) 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@73 -- # killprocess 735101 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@948 -- # '[' -z 735101 ']' 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@952 -- # kill -0 735101 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@953 -- # uname 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 735101 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@966 -- # echo 'killing process with pid 735101' 00:26:17.240 killing process with pid 735101 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@967 -- # kill 735101 00:26:17.240 Received shutdown signal, test time was about 2.000000 seconds 00:26:17.240 00:26:17.240 Latency(us) 00:26:17.240 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:17.240 =================================================================================================================== 00:26:17.240 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@972 -- # wait 735101 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- host/digest.sh@116 -- # killprocess 732859 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@948 -- # '[' -z 732859 ']' 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@952 -- # kill -0 732859 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@953 -- # uname 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 732859 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@966 -- # echo 'killing process with pid 732859' 00:26:17.240 killing process with pid 732859 00:26:17.240 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@967 -- # kill 732859 00:26:17.499 18:12:10 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@972 -- # wait 732859 00:26:17.499 00:26:17.499 real 0m16.833s 00:26:17.499 user 0m32.382s 00:26:17.499 sys 0m4.289s 00:26:17.499 18:12:11 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@1124 -- # xtrace_disable 00:26:17.499 18:12:11 nvmf_tcp.nvmf_digest.nvmf_digest_error -- common/autotest_common.sh@10 -- # set +x 00:26:17.499 ************************************ 00:26:17.499 END TEST nvmf_digest_error 00:26:17.499 ************************************ 00:26:17.499 18:12:11 nvmf_tcp.nvmf_digest -- common/autotest_common.sh@1142 -- # return 0 00:26:17.499 18:12:11 nvmf_tcp.nvmf_digest -- host/digest.sh@149 -- # trap - SIGINT SIGTERM EXIT 00:26:17.499 18:12:11 nvmf_tcp.nvmf_digest -- host/digest.sh@150 -- # nvmftestfini 00:26:17.499 18:12:11 nvmf_tcp.nvmf_digest -- nvmf/common.sh@488 -- # nvmfcleanup 00:26:17.499 18:12:11 nvmf_tcp.nvmf_digest -- nvmf/common.sh@117 -- # sync 00:26:17.499 18:12:11 nvmf_tcp.nvmf_digest -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:26:17.499 18:12:11 nvmf_tcp.nvmf_digest -- nvmf/common.sh@120 -- # set +e 00:26:17.499 18:12:11 nvmf_tcp.nvmf_digest -- nvmf/common.sh@121 -- # for i in {1..20} 00:26:17.499 18:12:11 nvmf_tcp.nvmf_digest -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:26:17.499 rmmod nvme_tcp 00:26:17.499 rmmod nvme_fabrics 00:26:17.758 rmmod nvme_keyring 00:26:17.758 18:12:11 nvmf_tcp.nvmf_digest -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:26:17.758 18:12:11 nvmf_tcp.nvmf_digest -- nvmf/common.sh@124 -- # set -e 00:26:17.758 18:12:11 nvmf_tcp.nvmf_digest -- nvmf/common.sh@125 -- # return 0 00:26:17.758 18:12:11 nvmf_tcp.nvmf_digest -- nvmf/common.sh@489 -- # '[' -n 732859 ']' 00:26:17.758 18:12:11 nvmf_tcp.nvmf_digest -- nvmf/common.sh@490 -- # killprocess 732859 00:26:17.758 18:12:11 nvmf_tcp.nvmf_digest -- common/autotest_common.sh@948 -- # '[' -z 732859 ']' 00:26:17.758 18:12:11 nvmf_tcp.nvmf_digest -- common/autotest_common.sh@952 -- # kill -0 732859 00:26:17.758 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 952: kill: (732859) - No such process 00:26:17.758 18:12:11 nvmf_tcp.nvmf_digest -- common/autotest_common.sh@975 -- # echo 'Process with pid 732859 is not found' 00:26:17.758 Process with pid 732859 is not found 00:26:17.758 18:12:11 nvmf_tcp.nvmf_digest -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:26:17.758 18:12:11 nvmf_tcp.nvmf_digest -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:26:17.758 18:12:11 nvmf_tcp.nvmf_digest -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:26:17.758 18:12:11 nvmf_tcp.nvmf_digest -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:26:17.758 18:12:11 nvmf_tcp.nvmf_digest -- nvmf/common.sh@278 -- # remove_spdk_ns 00:26:17.758 18:12:11 nvmf_tcp.nvmf_digest -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:17.758 18:12:11 nvmf_tcp.nvmf_digest -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:26:17.758 18:12:11 nvmf_tcp.nvmf_digest -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:19.678 18:12:13 nvmf_tcp.nvmf_digest -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:26:19.678 00:26:19.678 real 0m41.623s 00:26:19.678 user 1m6.216s 00:26:19.678 sys 0m13.071s 00:26:19.678 18:12:13 nvmf_tcp.nvmf_digest -- common/autotest_common.sh@1124 -- # xtrace_disable 00:26:19.678 18:12:13 nvmf_tcp.nvmf_digest -- common/autotest_common.sh@10 -- # set +x 00:26:19.678 ************************************ 00:26:19.678 END TEST nvmf_digest 00:26:19.678 ************************************ 00:26:19.678 18:12:13 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:26:19.678 18:12:13 nvmf_tcp -- nvmf/nvmf.sh@111 -- # [[ 0 -eq 1 ]] 00:26:19.678 18:12:13 nvmf_tcp -- nvmf/nvmf.sh@116 -- # [[ 0 -eq 1 ]] 00:26:19.678 18:12:13 nvmf_tcp -- nvmf/nvmf.sh@121 -- # [[ phy == phy ]] 00:26:19.678 18:12:13 nvmf_tcp -- nvmf/nvmf.sh@122 -- # run_test nvmf_bdevperf /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/bdevperf.sh --transport=tcp 00:26:19.678 18:12:13 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:26:19.678 18:12:13 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:26:19.678 18:12:13 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:26:19.678 ************************************ 00:26:19.678 START TEST nvmf_bdevperf 00:26:19.678 ************************************ 00:26:19.678 18:12:13 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/bdevperf.sh --transport=tcp 00:26:19.938 * Looking for test storage... 00:26:19.938 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@7 -- # uname -s 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- paths/export.sh@5 -- # export PATH 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@47 -- # : 0 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@51 -- # have_pci_nics=0 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@11 -- # MALLOC_BDEV_SIZE=64 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@12 -- # MALLOC_BLOCK_SIZE=512 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@24 -- # nvmftestinit 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@448 -- # prepare_net_devs 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@410 -- # local -g is_hw=no 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@412 -- # remove_spdk_ns 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@285 -- # xtrace_disable 00:26:19.938 18:12:13 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@291 -- # pci_devs=() 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@291 -- # local -a pci_devs 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@292 -- # pci_net_devs=() 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@293 -- # pci_drivers=() 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@293 -- # local -A pci_drivers 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@295 -- # net_devs=() 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@295 -- # local -ga net_devs 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@296 -- # e810=() 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@296 -- # local -ga e810 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@297 -- # x722=() 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@297 -- # local -ga x722 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@298 -- # mlx=() 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@298 -- # local -ga mlx 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:26:25.210 Found 0000:86:00.0 (0x8086 - 0x159b) 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:26:25.210 Found 0000:86:00.1 (0x8086 - 0x159b) 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@390 -- # [[ up == up ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:26:25.210 Found net devices under 0000:86:00.0: cvl_0_0 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@390 -- # [[ up == up ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:26:25.210 Found net devices under 0000:86:00.1: cvl_0_1 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@414 -- # is_hw=yes 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:26:25.210 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:26:25.211 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:26:25.211 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:26:25.211 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:26:25.211 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:26:25.211 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:26:25.211 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:26:25.211 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:26:25.211 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:26:25.211 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:26:25.211 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:26:25.211 18:12:17 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:26:25.211 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:26:25.211 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.164 ms 00:26:25.211 00:26:25.211 --- 10.0.0.2 ping statistics --- 00:26:25.211 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:25.211 rtt min/avg/max/mdev = 0.164/0.164/0.164/0.000 ms 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:26:25.211 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:26:25.211 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.105 ms 00:26:25.211 00:26:25.211 --- 10.0.0.1 ping statistics --- 00:26:25.211 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:25.211 rtt min/avg/max/mdev = 0.105/0.105/0.105/0.000 ms 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@422 -- # return 0 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@25 -- # tgt_init 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@15 -- # nvmfappstart -m 0xE 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@722 -- # xtrace_disable 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@481 -- # nvmfpid=739498 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@482 -- # waitforlisten 739498 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@829 -- # '[' -z 739498 ']' 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@834 -- # local max_retries=100 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:25.211 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@838 -- # xtrace_disable 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:26:25.211 18:12:18 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:26:25.211 [2024-07-15 18:12:18.270328] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:26:25.211 [2024-07-15 18:12:18.270371] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:25.211 EAL: No free 2048 kB hugepages reported on node 1 00:26:25.211 [2024-07-15 18:12:18.326714] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:26:25.211 [2024-07-15 18:12:18.406265] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:26:25.211 [2024-07-15 18:12:18.406301] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:26:25.211 [2024-07-15 18:12:18.406308] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:26:25.211 [2024-07-15 18:12:18.406314] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:26:25.211 [2024-07-15 18:12:18.406319] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:26:25.211 [2024-07-15 18:12:18.406356] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:26:25.211 [2024-07-15 18:12:18.406477] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:26:25.211 [2024-07-15 18:12:18.406478] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@862 -- # return 0 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@728 -- # xtrace_disable 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:26:25.470 [2024-07-15 18:12:19.110007] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@18 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:26:25.470 Malloc0 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@19 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@21 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:26:25.470 [2024-07-15 18:12:19.170111] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@27 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/62 -q 128 -o 4096 -w verify -t 1 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@27 -- # gen_nvmf_target_json 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@532 -- # config=() 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@532 -- # local subsystem config 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:26:25.470 18:12:19 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:26:25.470 { 00:26:25.470 "params": { 00:26:25.470 "name": "Nvme$subsystem", 00:26:25.470 "trtype": "$TEST_TRANSPORT", 00:26:25.470 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:25.470 "adrfam": "ipv4", 00:26:25.470 "trsvcid": "$NVMF_PORT", 00:26:25.471 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:25.471 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:25.471 "hdgst": ${hdgst:-false}, 00:26:25.471 "ddgst": ${ddgst:-false} 00:26:25.471 }, 00:26:25.471 "method": "bdev_nvme_attach_controller" 00:26:25.471 } 00:26:25.471 EOF 00:26:25.471 )") 00:26:25.471 18:12:19 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@554 -- # cat 00:26:25.471 18:12:19 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@556 -- # jq . 00:26:25.471 18:12:19 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@557 -- # IFS=, 00:26:25.471 18:12:19 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:26:25.471 "params": { 00:26:25.471 "name": "Nvme1", 00:26:25.471 "trtype": "tcp", 00:26:25.471 "traddr": "10.0.0.2", 00:26:25.471 "adrfam": "ipv4", 00:26:25.471 "trsvcid": "4420", 00:26:25.471 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:26:25.471 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:26:25.471 "hdgst": false, 00:26:25.471 "ddgst": false 00:26:25.471 }, 00:26:25.471 "method": "bdev_nvme_attach_controller" 00:26:25.471 }' 00:26:25.730 [2024-07-15 18:12:19.220087] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:26:25.730 [2024-07-15 18:12:19.220131] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid739743 ] 00:26:25.730 EAL: No free 2048 kB hugepages reported on node 1 00:26:25.730 [2024-07-15 18:12:19.273699] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:25.730 [2024-07-15 18:12:19.346709] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:26:25.989 Running I/O for 1 seconds... 00:26:27.362 00:26:27.362 Latency(us) 00:26:27.362 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:27.362 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:26:27.362 Verification LBA range: start 0x0 length 0x4000 00:26:27.362 Nvme1n1 : 1.01 11076.15 43.27 0.00 0.00 11513.35 2564.45 12480.33 00:26:27.362 =================================================================================================================== 00:26:27.362 Total : 11076.15 43.27 0.00 0.00 11513.35 2564.45 12480.33 00:26:27.362 18:12:20 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@30 -- # bdevperfpid=739984 00:26:27.362 18:12:20 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@32 -- # sleep 3 00:26:27.362 18:12:20 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf --json /dev/fd/63 -q 128 -o 4096 -w verify -t 15 -f 00:26:27.362 18:12:20 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@29 -- # gen_nvmf_target_json 00:26:27.362 18:12:20 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@532 -- # config=() 00:26:27.362 18:12:20 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@532 -- # local subsystem config 00:26:27.362 18:12:20 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:26:27.362 18:12:20 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:26:27.362 { 00:26:27.362 "params": { 00:26:27.362 "name": "Nvme$subsystem", 00:26:27.362 "trtype": "$TEST_TRANSPORT", 00:26:27.362 "traddr": "$NVMF_FIRST_TARGET_IP", 00:26:27.362 "adrfam": "ipv4", 00:26:27.362 "trsvcid": "$NVMF_PORT", 00:26:27.362 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:26:27.362 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:26:27.362 "hdgst": ${hdgst:-false}, 00:26:27.362 "ddgst": ${ddgst:-false} 00:26:27.362 }, 00:26:27.362 "method": "bdev_nvme_attach_controller" 00:26:27.362 } 00:26:27.362 EOF 00:26:27.362 )") 00:26:27.362 18:12:20 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@554 -- # cat 00:26:27.362 18:12:20 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@556 -- # jq . 00:26:27.362 18:12:20 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@557 -- # IFS=, 00:26:27.362 18:12:20 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:26:27.362 "params": { 00:26:27.362 "name": "Nvme1", 00:26:27.362 "trtype": "tcp", 00:26:27.362 "traddr": "10.0.0.2", 00:26:27.362 "adrfam": "ipv4", 00:26:27.362 "trsvcid": "4420", 00:26:27.362 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:26:27.362 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:26:27.362 "hdgst": false, 00:26:27.362 "ddgst": false 00:26:27.362 }, 00:26:27.362 "method": "bdev_nvme_attach_controller" 00:26:27.362 }' 00:26:27.362 [2024-07-15 18:12:20.939102] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:26:27.362 [2024-07-15 18:12:20.939153] [ DPDK EAL parameters: bdevperf --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid739984 ] 00:26:27.362 EAL: No free 2048 kB hugepages reported on node 1 00:26:27.362 [2024-07-15 18:12:20.993281] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:26:27.362 [2024-07-15 18:12:21.062871] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:26:27.639 Running I/O for 15 seconds... 00:26:30.934 18:12:23 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@33 -- # kill -9 739498 00:26:30.934 18:12:23 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@35 -- # sleep 3 00:26:30.934 [2024-07-15 18:12:23.911523] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:95 nsid:1 lba:104744 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911563] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911582] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:49 nsid:1 lba:104752 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911591] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911601] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:9 nsid:1 lba:104760 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911608] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911616] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:88 nsid:1 lba:104768 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911623] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911632] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:64 nsid:1 lba:104776 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911639] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911647] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:38 nsid:1 lba:104784 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911653] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911662] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:5 nsid:1 lba:104792 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911669] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911678] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:17 nsid:1 lba:104800 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911685] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911694] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:126 nsid:1 lba:104808 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911707] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911716] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:42 nsid:1 lba:104816 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911723] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911732] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:97 nsid:1 lba:104824 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911739] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911747] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:86 nsid:1 lba:104832 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911754] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911762] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:55 nsid:1 lba:104840 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911771] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911781] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:102 nsid:1 lba:104848 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911789] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911797] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:103 nsid:1 lba:104856 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911804] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911814] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:56 nsid:1 lba:104864 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911822] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911831] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:92 nsid:1 lba:104872 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911839] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911850] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:89 nsid:1 lba:104880 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911858] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911866] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:2 nsid:1 lba:104888 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911873] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911883] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:119 nsid:1 lba:104896 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911889] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911898] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:114 nsid:1 lba:104904 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911904] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911914] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:8 nsid:1 lba:104912 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911921] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911929] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:120 nsid:1 lba:104920 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911936] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911943] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:105 nsid:1 lba:104928 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911950] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911958] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:34 nsid:1 lba:104936 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911964] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911972] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:14 nsid:1 lba:104944 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911978] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.911986] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:4 nsid:1 lba:104952 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.911993] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912000] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:70 nsid:1 lba:104960 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912009] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912018] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:106 nsid:1 lba:104968 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912024] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912032] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:107 nsid:1 lba:104976 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912039] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912046] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:45 nsid:1 lba:104984 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912053] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912060] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:99 nsid:1 lba:104992 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912067] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912074] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:22 nsid:1 lba:105000 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912081] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912094] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:12 nsid:1 lba:105008 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912101] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912109] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:47 nsid:1 lba:105016 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912116] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912124] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:123 nsid:1 lba:105024 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912130] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912138] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:104 nsid:1 lba:105032 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912144] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912152] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:94 nsid:1 lba:105040 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912159] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912167] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:52 nsid:1 lba:105048 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912174] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912182] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:110 nsid:1 lba:105752 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:30.934 [2024-07-15 18:12:23.912188] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912196] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:50 nsid:1 lba:105056 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912202] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912210] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:59 nsid:1 lba:105064 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912216] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912350] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:60 nsid:1 lba:105072 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912358] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912367] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:68 nsid:1 lba:105080 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912374] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912382] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:16 nsid:1 lba:105088 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912388] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912396] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:98 nsid:1 lba:105096 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912403] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912412] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:28 nsid:1 lba:105104 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912418] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912426] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:81 nsid:1 lba:105112 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912432] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912440] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:25 nsid:1 lba:105120 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912446] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912456] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:100 nsid:1 lba:105128 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912462] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.934 [2024-07-15 18:12:23.912470] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:18 nsid:1 lba:105136 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.934 [2024-07-15 18:12:23.912476] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912484] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:76 nsid:1 lba:105144 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912490] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912497] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:11 nsid:1 lba:105152 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912503] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912511] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:69 nsid:1 lba:105160 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912518] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912525] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:85 nsid:1 lba:105168 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912531] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: WRITE sqid:1 cid:84 nsid:1 lba:105760 len:8 SGL DATA BLOCK OFFSET 0x0 len:0x1000 00:26:30.935 [2024-07-15 18:12:23.912545] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912553] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:20 nsid:1 lba:105176 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912560] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912568] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:111 nsid:1 lba:105184 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912582] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:71 nsid:1 lba:105192 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912588] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912597] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:72 nsid:1 lba:105200 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912604] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912612] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:61 nsid:1 lba:105208 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912618] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912627] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:74 nsid:1 lba:105216 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912633] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912641] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:15 nsid:1 lba:105224 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912647] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912655] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:79 nsid:1 lba:105232 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912661] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912669] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:101 nsid:1 lba:105240 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912675] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912684] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:87 nsid:1 lba:105248 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912690] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912698] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:1 nsid:1 lba:105256 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912704] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912712] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:48 nsid:1 lba:105264 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912719] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912727] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:105272 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912733] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912741] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:33 nsid:1 lba:105280 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912747] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912755] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:63 nsid:1 lba:105288 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912761] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912769] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:113 nsid:1 lba:105296 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912777] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912785] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:41 nsid:1 lba:105304 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912791] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912799] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:117 nsid:1 lba:105312 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912805] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912812] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:93 nsid:1 lba:105320 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912818] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912826] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:67 nsid:1 lba:105328 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912833] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912840] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:82 nsid:1 lba:105336 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912846] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912854] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:3 nsid:1 lba:105344 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912860] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912869] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:46 nsid:1 lba:105352 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912875] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912883] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:116 nsid:1 lba:105360 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912890] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912898] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:78 nsid:1 lba:105368 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912904] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912912] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:37 nsid:1 lba:105376 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912918] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912926] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:35 nsid:1 lba:105384 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912933] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912940] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:96 nsid:1 lba:105392 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912947] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912955] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:19 nsid:1 lba:105400 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912961] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912969] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:32 nsid:1 lba:105408 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912975] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912983] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:13 nsid:1 lba:105416 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.912989] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.912998] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:24 nsid:1 lba:105424 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913004] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913012] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:31 nsid:1 lba:105432 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913018] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913027] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:27 nsid:1 lba:105440 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913033] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913041] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:51 nsid:1 lba:105448 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913048] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913055] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:90 nsid:1 lba:105456 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913062] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913069] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:58 nsid:1 lba:105464 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913076] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913084] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:53 nsid:1 lba:105472 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913090] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913098] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:91 nsid:1 lba:105480 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913104] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913113] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:73 nsid:1 lba:105488 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913119] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913127] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:43 nsid:1 lba:105496 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913135] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913143] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:125 nsid:1 lba:105504 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913150] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913158] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:54 nsid:1 lba:105512 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913164] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913172] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:7 nsid:1 lba:105520 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913178] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913188] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:66 nsid:1 lba:105528 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913194] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913202] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:39 nsid:1 lba:105536 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913208] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913215] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:30 nsid:1 lba:105544 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913221] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913233] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:112 nsid:1 lba:105552 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913239] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913247] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:57 nsid:1 lba:105560 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913254] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913262] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:108 nsid:1 lba:105568 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913268] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913276] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:83 nsid:1 lba:105576 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913282] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913290] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:29 nsid:1 lba:105584 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913297] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913305] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:75 nsid:1 lba:105592 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913311] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913323] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:10 nsid:1 lba:105600 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913329] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913337] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:115 nsid:1 lba:105608 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913343] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913351] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:109 nsid:1 lba:105616 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913358] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913366] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:36 nsid:1 lba:105624 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913372] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913381] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:77 nsid:1 lba:105632 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.935 [2024-07-15 18:12:23.913388] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.935 [2024-07-15 18:12:23.913395] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:62 nsid:1 lba:105640 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.936 [2024-07-15 18:12:23.913402] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.936 [2024-07-15 18:12:23.913410] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:121 nsid:1 lba:105648 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.936 [2024-07-15 18:12:23.913416] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.936 [2024-07-15 18:12:23.913425] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:80 nsid:1 lba:105656 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.936 [2024-07-15 18:12:23.913432] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.936 [2024-07-15 18:12:23.913439] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:118 nsid:1 lba:105664 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.936 [2024-07-15 18:12:23.913445] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.936 [2024-07-15 18:12:23.913453] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:44 nsid:1 lba:105672 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.936 [2024-07-15 18:12:23.913460] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.936 [2024-07-15 18:12:23.913467] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:26 nsid:1 lba:105680 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.936 [2024-07-15 18:12:23.913473] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.936 [2024-07-15 18:12:23.913481] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:40 nsid:1 lba:105688 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.936 [2024-07-15 18:12:23.913487] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.936 [2024-07-15 18:12:23.913495] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:6 nsid:1 lba:105696 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.936 [2024-07-15 18:12:23.913503] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.936 [2024-07-15 18:12:23.913511] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:23 nsid:1 lba:105704 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.936 [2024-07-15 18:12:23.913517] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.936 [2024-07-15 18:12:23.913525] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:124 nsid:1 lba:105712 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.936 [2024-07-15 18:12:23.913531] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.936 [2024-07-15 18:12:23.913539] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:21 nsid:1 lba:105720 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.936 [2024-07-15 18:12:23.913545] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.936 [2024-07-15 18:12:23.913553] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:65 nsid:1 lba:105728 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.936 [2024-07-15 18:12:23.913559] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.936 [2024-07-15 18:12:23.913567] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:122 nsid:1 lba:105736 len:8 SGL TRANSPORT DATA BLOCK TRANSPORT 0x0 00:26:30.936 [2024-07-15 18:12:23.913574] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.936 [2024-07-15 18:12:23.913581] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1784c70 is same with the state(5) to be set 00:26:30.936 [2024-07-15 18:12:23.913589] nvme_qpair.c: 579:nvme_qpair_abort_queued_reqs: *ERROR*: aborting queued i/o 00:26:30.936 [2024-07-15 18:12:23.913594] nvme_qpair.c: 558:nvme_qpair_manual_complete_request: *NOTICE*: Command completed manually: 00:26:30.936 [2024-07-15 18:12:23.913600] nvme_qpair.c: 243:nvme_io_qpair_print_command: *NOTICE*: READ sqid:1 cid:0 nsid:1 lba:105744 len:8 PRP1 0x0 PRP2 0x0 00:26:30.936 [2024-07-15 18:12:23.913607] nvme_qpair.c: 474:spdk_nvme_print_completion: *NOTICE*: ABORTED - SQ DELETION (00/08) qid:1 cid:0 cdw0:0 sqhd:0000 p:0 m:0 dnr:0 00:26:30.936 [2024-07-15 18:12:23.913650] bdev_nvme.c:1612:bdev_nvme_disconnected_qpair_cb: *NOTICE*: qpair 0x1784c70 was disconnected and freed. reset controller. 00:26:30.936 [2024-07-15 18:12:23.916774] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.936 [2024-07-15 18:12:23.916825] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.936 [2024-07-15 18:12:23.917476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.936 [2024-07-15 18:12:23.917491] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.936 [2024-07-15 18:12:23.917499] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.936 [2024-07-15 18:12:23.917676] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.936 [2024-07-15 18:12:23.917852] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.936 [2024-07-15 18:12:23.917860] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.936 [2024-07-15 18:12:23.917867] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.936 [2024-07-15 18:12:23.920697] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.936 [2024-07-15 18:12:23.930032] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.936 [2024-07-15 18:12:23.930524] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.936 [2024-07-15 18:12:23.930568] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.936 [2024-07-15 18:12:23.930590] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.936 [2024-07-15 18:12:23.931167] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.936 [2024-07-15 18:12:23.931645] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.936 [2024-07-15 18:12:23.931653] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.936 [2024-07-15 18:12:23.931660] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.936 [2024-07-15 18:12:23.934447] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.936 [2024-07-15 18:12:23.942927] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.936 [2024-07-15 18:12:23.943362] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.936 [2024-07-15 18:12:23.943400] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.936 [2024-07-15 18:12:23.943422] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.936 [2024-07-15 18:12:23.944001] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.936 [2024-07-15 18:12:23.944605] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.936 [2024-07-15 18:12:23.944631] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.936 [2024-07-15 18:12:23.944651] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.936 [2024-07-15 18:12:23.947366] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.936 [2024-07-15 18:12:23.955752] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.936 [2024-07-15 18:12:23.956234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.936 [2024-07-15 18:12:23.956250] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.936 [2024-07-15 18:12:23.956257] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.936 [2024-07-15 18:12:23.956428] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.936 [2024-07-15 18:12:23.956599] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.936 [2024-07-15 18:12:23.956606] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.936 [2024-07-15 18:12:23.956612] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.936 [2024-07-15 18:12:23.959312] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.936 [2024-07-15 18:12:23.968662] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.936 [2024-07-15 18:12:23.969166] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.936 [2024-07-15 18:12:23.969208] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.936 [2024-07-15 18:12:23.969244] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.936 [2024-07-15 18:12:23.969779] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.936 [2024-07-15 18:12:23.969956] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.936 [2024-07-15 18:12:23.969963] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.936 [2024-07-15 18:12:23.969969] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.936 [2024-07-15 18:12:23.972651] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.936 [2024-07-15 18:12:23.981517] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.936 [2024-07-15 18:12:23.981961] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.936 [2024-07-15 18:12:23.982003] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.936 [2024-07-15 18:12:23.982025] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.936 [2024-07-15 18:12:23.982530] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.936 [2024-07-15 18:12:23.982702] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.936 [2024-07-15 18:12:23.982709] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.936 [2024-07-15 18:12:23.982715] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.936 [2024-07-15 18:12:23.985384] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.936 [2024-07-15 18:12:23.994371] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.936 [2024-07-15 18:12:23.994856] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.936 [2024-07-15 18:12:23.994898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.936 [2024-07-15 18:12:23.994919] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.936 [2024-07-15 18:12:23.995512] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.936 [2024-07-15 18:12:23.995759] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.936 [2024-07-15 18:12:23.995766] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.936 [2024-07-15 18:12:23.995772] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.936 [2024-07-15 18:12:23.998405] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.936 [2024-07-15 18:12:24.007195] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.936 [2024-07-15 18:12:24.007673] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.936 [2024-07-15 18:12:24.007689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.936 [2024-07-15 18:12:24.007695] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.936 [2024-07-15 18:12:24.007856] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.936 [2024-07-15 18:12:24.008018] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.936 [2024-07-15 18:12:24.008026] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.936 [2024-07-15 18:12:24.008034] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.936 [2024-07-15 18:12:24.010721] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.936 [2024-07-15 18:12:24.020074] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.936 [2024-07-15 18:12:24.020547] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.936 [2024-07-15 18:12:24.020563] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.936 [2024-07-15 18:12:24.020569] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.936 [2024-07-15 18:12:24.020740] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.936 [2024-07-15 18:12:24.020915] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.936 [2024-07-15 18:12:24.020922] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.936 [2024-07-15 18:12:24.020928] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.936 [2024-07-15 18:12:24.023663] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.936 [2024-07-15 18:12:24.032996] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.936 [2024-07-15 18:12:24.033407] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.936 [2024-07-15 18:12:24.033450] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.936 [2024-07-15 18:12:24.033471] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.936 [2024-07-15 18:12:24.034020] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.936 [2024-07-15 18:12:24.034191] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.936 [2024-07-15 18:12:24.034199] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.936 [2024-07-15 18:12:24.034205] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.936 [2024-07-15 18:12:24.036944] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.936 [2024-07-15 18:12:24.045813] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.936 [2024-07-15 18:12:24.046220] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.936 [2024-07-15 18:12:24.046282] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.936 [2024-07-15 18:12:24.046303] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.936 [2024-07-15 18:12:24.046758] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.936 [2024-07-15 18:12:24.046929] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.936 [2024-07-15 18:12:24.046937] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.936 [2024-07-15 18:12:24.046943] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.937 [2024-07-15 18:12:24.049625] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.937 [2024-07-15 18:12:24.058611] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.937 [2024-07-15 18:12:24.059058] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.937 [2024-07-15 18:12:24.059076] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.937 [2024-07-15 18:12:24.059082] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.937 [2024-07-15 18:12:24.059249] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.937 [2024-07-15 18:12:24.059437] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.937 [2024-07-15 18:12:24.059445] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.937 [2024-07-15 18:12:24.059450] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.937 [2024-07-15 18:12:24.062117] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.937 [2024-07-15 18:12:24.071517] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.937 [2024-07-15 18:12:24.071990] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.937 [2024-07-15 18:12:24.072006] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.937 [2024-07-15 18:12:24.072012] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.937 [2024-07-15 18:12:24.072183] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.937 [2024-07-15 18:12:24.072359] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.937 [2024-07-15 18:12:24.072372] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.937 [2024-07-15 18:12:24.072379] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.937 [2024-07-15 18:12:24.075116] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.937 [2024-07-15 18:12:24.084438] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.937 [2024-07-15 18:12:24.084916] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.937 [2024-07-15 18:12:24.084931] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.937 [2024-07-15 18:12:24.084938] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.937 [2024-07-15 18:12:24.085108] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.937 [2024-07-15 18:12:24.085285] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.937 [2024-07-15 18:12:24.085293] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.937 [2024-07-15 18:12:24.085299] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.937 [2024-07-15 18:12:24.087965] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.937 [2024-07-15 18:12:24.097271] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.937 [2024-07-15 18:12:24.097695] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.937 [2024-07-15 18:12:24.097710] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.937 [2024-07-15 18:12:24.097717] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.937 [2024-07-15 18:12:24.097878] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.937 [2024-07-15 18:12:24.098043] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.937 [2024-07-15 18:12:24.098050] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.937 [2024-07-15 18:12:24.098055] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.937 [2024-07-15 18:12:24.100738] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.937 [2024-07-15 18:12:24.110186] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.937 [2024-07-15 18:12:24.110663] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.937 [2024-07-15 18:12:24.110679] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.937 [2024-07-15 18:12:24.110686] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.937 [2024-07-15 18:12:24.110856] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.937 [2024-07-15 18:12:24.111027] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.937 [2024-07-15 18:12:24.111035] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.937 [2024-07-15 18:12:24.111041] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.937 [2024-07-15 18:12:24.113716] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.937 [2024-07-15 18:12:24.123068] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.937 [2024-07-15 18:12:24.123541] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.937 [2024-07-15 18:12:24.123556] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.937 [2024-07-15 18:12:24.123563] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.937 [2024-07-15 18:12:24.123733] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.937 [2024-07-15 18:12:24.123903] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.937 [2024-07-15 18:12:24.123911] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.937 [2024-07-15 18:12:24.123917] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.937 [2024-07-15 18:12:24.126632] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.937 [2024-07-15 18:12:24.136065] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.937 [2024-07-15 18:12:24.136522] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.937 [2024-07-15 18:12:24.136538] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.937 [2024-07-15 18:12:24.136544] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.937 [2024-07-15 18:12:24.136714] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.937 [2024-07-15 18:12:24.136884] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.937 [2024-07-15 18:12:24.136892] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.937 [2024-07-15 18:12:24.136897] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.937 [2024-07-15 18:12:24.139577] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.937 [2024-07-15 18:12:24.148900] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.937 [2024-07-15 18:12:24.149363] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.937 [2024-07-15 18:12:24.149405] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.937 [2024-07-15 18:12:24.149426] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.937 [2024-07-15 18:12:24.150005] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.937 [2024-07-15 18:12:24.150453] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.937 [2024-07-15 18:12:24.150464] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.937 [2024-07-15 18:12:24.150473] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.937 [2024-07-15 18:12:24.154517] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.937 [2024-07-15 18:12:24.162436] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.937 [2024-07-15 18:12:24.162905] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.937 [2024-07-15 18:12:24.162921] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.937 [2024-07-15 18:12:24.162928] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.937 [2024-07-15 18:12:24.163103] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.937 [2024-07-15 18:12:24.163286] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.937 [2024-07-15 18:12:24.163295] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.937 [2024-07-15 18:12:24.163301] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.937 [2024-07-15 18:12:24.166122] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.937 [2024-07-15 18:12:24.175464] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.937 [2024-07-15 18:12:24.175906] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.937 [2024-07-15 18:12:24.175922] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.937 [2024-07-15 18:12:24.175929] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.937 [2024-07-15 18:12:24.176104] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.937 [2024-07-15 18:12:24.176287] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.937 [2024-07-15 18:12:24.176295] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.937 [2024-07-15 18:12:24.176302] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.937 [2024-07-15 18:12:24.179121] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.937 [2024-07-15 18:12:24.188578] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.937 [2024-07-15 18:12:24.189040] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.937 [2024-07-15 18:12:24.189056] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.937 [2024-07-15 18:12:24.189066] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.937 [2024-07-15 18:12:24.189248] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.937 [2024-07-15 18:12:24.189425] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.937 [2024-07-15 18:12:24.189433] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.937 [2024-07-15 18:12:24.189439] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.937 [2024-07-15 18:12:24.192211] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.937 [2024-07-15 18:12:24.201623] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.937 [2024-07-15 18:12:24.202078] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.937 [2024-07-15 18:12:24.202094] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.937 [2024-07-15 18:12:24.202101] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.937 [2024-07-15 18:12:24.202277] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.937 [2024-07-15 18:12:24.202448] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.937 [2024-07-15 18:12:24.202456] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.937 [2024-07-15 18:12:24.202462] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.937 [2024-07-15 18:12:24.205129] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.937 [2024-07-15 18:12:24.214472] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.937 [2024-07-15 18:12:24.214924] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.937 [2024-07-15 18:12:24.214940] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.937 [2024-07-15 18:12:24.214947] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.937 [2024-07-15 18:12:24.215119] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.937 [2024-07-15 18:12:24.215297] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.937 [2024-07-15 18:12:24.215305] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.937 [2024-07-15 18:12:24.215311] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.937 [2024-07-15 18:12:24.217979] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.937 [2024-07-15 18:12:24.227388] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.937 [2024-07-15 18:12:24.227819] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.937 [2024-07-15 18:12:24.227834] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.937 [2024-07-15 18:12:24.227840] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.937 [2024-07-15 18:12:24.228002] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.937 [2024-07-15 18:12:24.228163] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.937 [2024-07-15 18:12:24.228173] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.937 [2024-07-15 18:12:24.228179] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.937 [2024-07-15 18:12:24.230868] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.937 [2024-07-15 18:12:24.240179] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.937 [2024-07-15 18:12:24.240645] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.937 [2024-07-15 18:12:24.240686] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.937 [2024-07-15 18:12:24.240708] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.937 [2024-07-15 18:12:24.241250] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.937 [2024-07-15 18:12:24.241422] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.937 [2024-07-15 18:12:24.241444] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.937 [2024-07-15 18:12:24.241453] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.937 [2024-07-15 18:12:24.245496] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.937 [2024-07-15 18:12:24.253519] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.937 [2024-07-15 18:12:24.253983] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.937 [2024-07-15 18:12:24.254025] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.937 [2024-07-15 18:12:24.254046] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.937 [2024-07-15 18:12:24.254537] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.937 [2024-07-15 18:12:24.254708] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.937 [2024-07-15 18:12:24.254716] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.937 [2024-07-15 18:12:24.254722] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.937 [2024-07-15 18:12:24.257437] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.937 [2024-07-15 18:12:24.266528] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.937 [2024-07-15 18:12:24.267025] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.937 [2024-07-15 18:12:24.267067] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.937 [2024-07-15 18:12:24.267088] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.937 [2024-07-15 18:12:24.267579] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.938 [2024-07-15 18:12:24.267752] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.938 [2024-07-15 18:12:24.267761] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.938 [2024-07-15 18:12:24.267766] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.938 [2024-07-15 18:12:24.270518] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.938 [2024-07-15 18:12:24.279475] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.938 [2024-07-15 18:12:24.279909] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.938 [2024-07-15 18:12:24.279947] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.938 [2024-07-15 18:12:24.279969] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.938 [2024-07-15 18:12:24.280538] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.938 [2024-07-15 18:12:24.280715] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.938 [2024-07-15 18:12:24.280724] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.938 [2024-07-15 18:12:24.280730] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.938 [2024-07-15 18:12:24.283415] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.938 [2024-07-15 18:12:24.292369] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.938 [2024-07-15 18:12:24.292856] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.938 [2024-07-15 18:12:24.292898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.938 [2024-07-15 18:12:24.292920] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.938 [2024-07-15 18:12:24.293392] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.938 [2024-07-15 18:12:24.293570] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.938 [2024-07-15 18:12:24.293578] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.938 [2024-07-15 18:12:24.293584] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.938 [2024-07-15 18:12:24.296310] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.938 [2024-07-15 18:12:24.305289] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.938 [2024-07-15 18:12:24.305604] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.938 [2024-07-15 18:12:24.305620] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.938 [2024-07-15 18:12:24.305627] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.938 [2024-07-15 18:12:24.305798] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.938 [2024-07-15 18:12:24.305971] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.938 [2024-07-15 18:12:24.305979] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.938 [2024-07-15 18:12:24.305985] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.938 [2024-07-15 18:12:24.308706] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.938 [2024-07-15 18:12:24.318435] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.938 [2024-07-15 18:12:24.318838] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.938 [2024-07-15 18:12:24.318880] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.938 [2024-07-15 18:12:24.318902] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.938 [2024-07-15 18:12:24.319499] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.938 [2024-07-15 18:12:24.319746] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.938 [2024-07-15 18:12:24.319757] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.938 [2024-07-15 18:12:24.319763] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.938 [2024-07-15 18:12:24.322592] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.938 [2024-07-15 18:12:24.331374] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.938 [2024-07-15 18:12:24.331745] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.938 [2024-07-15 18:12:24.331760] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.938 [2024-07-15 18:12:24.331767] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.938 [2024-07-15 18:12:24.331928] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.938 [2024-07-15 18:12:24.332090] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.938 [2024-07-15 18:12:24.332097] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.938 [2024-07-15 18:12:24.332103] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.938 [2024-07-15 18:12:24.334832] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.938 [2024-07-15 18:12:24.344328] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.938 [2024-07-15 18:12:24.344766] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.938 [2024-07-15 18:12:24.344782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.938 [2024-07-15 18:12:24.344788] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.938 [2024-07-15 18:12:24.344960] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.938 [2024-07-15 18:12:24.345132] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.938 [2024-07-15 18:12:24.345140] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.938 [2024-07-15 18:12:24.345146] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.938 [2024-07-15 18:12:24.347828] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.938 [2024-07-15 18:12:24.357432] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.938 [2024-07-15 18:12:24.357853] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.938 [2024-07-15 18:12:24.357869] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.938 [2024-07-15 18:12:24.357876] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.938 [2024-07-15 18:12:24.358046] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.938 [2024-07-15 18:12:24.358217] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.938 [2024-07-15 18:12:24.358232] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.938 [2024-07-15 18:12:24.358244] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.938 [2024-07-15 18:12:24.361023] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.938 [2024-07-15 18:12:24.370405] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.938 [2024-07-15 18:12:24.370814] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.938 [2024-07-15 18:12:24.370830] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.938 [2024-07-15 18:12:24.370837] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.938 [2024-07-15 18:12:24.371009] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.938 [2024-07-15 18:12:24.371181] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.938 [2024-07-15 18:12:24.371189] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.938 [2024-07-15 18:12:24.371194] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.938 [2024-07-15 18:12:24.373926] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.938 [2024-07-15 18:12:24.383275] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.938 [2024-07-15 18:12:24.383666] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.938 [2024-07-15 18:12:24.383681] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.938 [2024-07-15 18:12:24.383688] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.938 [2024-07-15 18:12:24.383859] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.938 [2024-07-15 18:12:24.384030] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.938 [2024-07-15 18:12:24.384038] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.938 [2024-07-15 18:12:24.384044] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.938 [2024-07-15 18:12:24.386724] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.938 [2024-07-15 18:12:24.396337] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.938 [2024-07-15 18:12:24.396650] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.938 [2024-07-15 18:12:24.396666] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.938 [2024-07-15 18:12:24.396673] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.938 [2024-07-15 18:12:24.396844] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.938 [2024-07-15 18:12:24.397015] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.938 [2024-07-15 18:12:24.397023] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.938 [2024-07-15 18:12:24.397029] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.938 [2024-07-15 18:12:24.399755] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.938 [2024-07-15 18:12:24.409314] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.938 [2024-07-15 18:12:24.409777] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.938 [2024-07-15 18:12:24.409792] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.938 [2024-07-15 18:12:24.409799] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.938 [2024-07-15 18:12:24.409976] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.938 [2024-07-15 18:12:24.410152] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.938 [2024-07-15 18:12:24.410161] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.938 [2024-07-15 18:12:24.410166] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.938 [2024-07-15 18:12:24.413002] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.938 [2024-07-15 18:12:24.422422] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.938 [2024-07-15 18:12:24.422899] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.938 [2024-07-15 18:12:24.422941] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.938 [2024-07-15 18:12:24.422962] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.938 [2024-07-15 18:12:24.423555] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.938 [2024-07-15 18:12:24.424016] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.938 [2024-07-15 18:12:24.424024] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.938 [2024-07-15 18:12:24.424031] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.938 [2024-07-15 18:12:24.426863] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.938 [2024-07-15 18:12:24.435555] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.938 [2024-07-15 18:12:24.435924] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.938 [2024-07-15 18:12:24.435940] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.938 [2024-07-15 18:12:24.435947] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.938 [2024-07-15 18:12:24.436123] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.938 [2024-07-15 18:12:24.436308] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.938 [2024-07-15 18:12:24.436316] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.938 [2024-07-15 18:12:24.436322] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.938 [2024-07-15 18:12:24.439147] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.938 [2024-07-15 18:12:24.448667] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.938 [2024-07-15 18:12:24.449005] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.938 [2024-07-15 18:12:24.449021] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.938 [2024-07-15 18:12:24.449028] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.938 [2024-07-15 18:12:24.449210] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.938 [2024-07-15 18:12:24.449401] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.938 [2024-07-15 18:12:24.449410] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.938 [2024-07-15 18:12:24.449416] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.938 [2024-07-15 18:12:24.452170] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.938 [2024-07-15 18:12:24.461667] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.938 [2024-07-15 18:12:24.462035] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.938 [2024-07-15 18:12:24.462051] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.938 [2024-07-15 18:12:24.462057] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.938 [2024-07-15 18:12:24.462236] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.938 [2024-07-15 18:12:24.462408] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.938 [2024-07-15 18:12:24.462416] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.938 [2024-07-15 18:12:24.462422] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.938 [2024-07-15 18:12:24.465097] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.938 [2024-07-15 18:12:24.474669] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.938 [2024-07-15 18:12:24.475014] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.938 [2024-07-15 18:12:24.475029] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.938 [2024-07-15 18:12:24.475036] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.938 [2024-07-15 18:12:24.475206] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.938 [2024-07-15 18:12:24.475405] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.938 [2024-07-15 18:12:24.475414] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.938 [2024-07-15 18:12:24.475420] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.938 [2024-07-15 18:12:24.478103] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.938 [2024-07-15 18:12:24.487609] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.938 [2024-07-15 18:12:24.487980] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.938 [2024-07-15 18:12:24.487996] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.938 [2024-07-15 18:12:24.488003] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.938 [2024-07-15 18:12:24.488174] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.938 [2024-07-15 18:12:24.488352] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.938 [2024-07-15 18:12:24.488360] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.938 [2024-07-15 18:12:24.488369] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.939 [2024-07-15 18:12:24.491109] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.939 [2024-07-15 18:12:24.500549] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.939 [2024-07-15 18:12:24.501030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.939 [2024-07-15 18:12:24.501046] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.939 [2024-07-15 18:12:24.501052] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.939 [2024-07-15 18:12:24.501223] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.939 [2024-07-15 18:12:24.501422] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.939 [2024-07-15 18:12:24.501430] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.939 [2024-07-15 18:12:24.501436] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.939 [2024-07-15 18:12:24.504147] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.939 [2024-07-15 18:12:24.513377] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.939 [2024-07-15 18:12:24.513749] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.939 [2024-07-15 18:12:24.513764] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.939 [2024-07-15 18:12:24.513771] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.939 [2024-07-15 18:12:24.513941] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.939 [2024-07-15 18:12:24.514112] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.939 [2024-07-15 18:12:24.514121] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.939 [2024-07-15 18:12:24.514127] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.939 [2024-07-15 18:12:24.516808] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.939 [2024-07-15 18:12:24.526550] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.939 [2024-07-15 18:12:24.526923] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.939 [2024-07-15 18:12:24.526939] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.939 [2024-07-15 18:12:24.526946] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.939 [2024-07-15 18:12:24.527122] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.939 [2024-07-15 18:12:24.527304] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.939 [2024-07-15 18:12:24.527312] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.939 [2024-07-15 18:12:24.527319] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.939 [2024-07-15 18:12:24.530137] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.939 [2024-07-15 18:12:24.539648] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.939 [2024-07-15 18:12:24.540014] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.939 [2024-07-15 18:12:24.540033] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.939 [2024-07-15 18:12:24.540040] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.939 [2024-07-15 18:12:24.540216] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.939 [2024-07-15 18:12:24.540398] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.939 [2024-07-15 18:12:24.540407] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.939 [2024-07-15 18:12:24.540413] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.939 [2024-07-15 18:12:24.543242] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.939 [2024-07-15 18:12:24.552761] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.939 [2024-07-15 18:12:24.553150] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.939 [2024-07-15 18:12:24.553166] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.939 [2024-07-15 18:12:24.553173] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.939 [2024-07-15 18:12:24.553355] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.939 [2024-07-15 18:12:24.553531] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.939 [2024-07-15 18:12:24.553539] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.939 [2024-07-15 18:12:24.553545] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.939 [2024-07-15 18:12:24.556369] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.939 [2024-07-15 18:12:24.565866] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.939 [2024-07-15 18:12:24.566331] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.939 [2024-07-15 18:12:24.566347] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.939 [2024-07-15 18:12:24.566354] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.939 [2024-07-15 18:12:24.566530] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.939 [2024-07-15 18:12:24.566706] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.939 [2024-07-15 18:12:24.566714] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.939 [2024-07-15 18:12:24.566720] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.939 [2024-07-15 18:12:24.569546] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.939 [2024-07-15 18:12:24.579059] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.939 [2024-07-15 18:12:24.579528] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.939 [2024-07-15 18:12:24.579544] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.939 [2024-07-15 18:12:24.579551] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.939 [2024-07-15 18:12:24.579727] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.939 [2024-07-15 18:12:24.579907] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.939 [2024-07-15 18:12:24.579914] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.939 [2024-07-15 18:12:24.579920] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.939 [2024-07-15 18:12:24.582748] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.939 [2024-07-15 18:12:24.592252] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.939 [2024-07-15 18:12:24.592721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.939 [2024-07-15 18:12:24.592737] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.939 [2024-07-15 18:12:24.592744] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.939 [2024-07-15 18:12:24.592920] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.939 [2024-07-15 18:12:24.593096] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.939 [2024-07-15 18:12:24.593104] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.939 [2024-07-15 18:12:24.593110] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.939 [2024-07-15 18:12:24.595932] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.939 [2024-07-15 18:12:24.605442] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.939 [2024-07-15 18:12:24.605910] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.939 [2024-07-15 18:12:24.605926] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.939 [2024-07-15 18:12:24.605933] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.939 [2024-07-15 18:12:24.606109] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.939 [2024-07-15 18:12:24.606292] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.939 [2024-07-15 18:12:24.606300] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.939 [2024-07-15 18:12:24.606306] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.939 [2024-07-15 18:12:24.609126] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.939 [2024-07-15 18:12:24.618633] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.939 [2024-07-15 18:12:24.619105] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.939 [2024-07-15 18:12:24.619121] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.939 [2024-07-15 18:12:24.619127] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.939 [2024-07-15 18:12:24.619308] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.939 [2024-07-15 18:12:24.619485] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.939 [2024-07-15 18:12:24.619492] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.939 [2024-07-15 18:12:24.619498] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.939 [2024-07-15 18:12:24.622322] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.939 [2024-07-15 18:12:24.631829] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.939 [2024-07-15 18:12:24.632293] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.939 [2024-07-15 18:12:24.632310] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.939 [2024-07-15 18:12:24.632317] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.939 [2024-07-15 18:12:24.632493] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.939 [2024-07-15 18:12:24.632669] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.939 [2024-07-15 18:12:24.632677] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.939 [2024-07-15 18:12:24.632683] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.939 [2024-07-15 18:12:24.635513] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:30.939 [2024-07-15 18:12:24.645016] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:30.939 [2024-07-15 18:12:24.645408] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:30.939 [2024-07-15 18:12:24.645424] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:30.939 [2024-07-15 18:12:24.645430] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:30.939 [2024-07-15 18:12:24.645606] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:30.939 [2024-07-15 18:12:24.645782] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:30.939 [2024-07-15 18:12:24.645790] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:30.939 [2024-07-15 18:12:24.645796] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:30.939 [2024-07-15 18:12:24.648623] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.199 [2024-07-15 18:12:24.658134] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.199 [2024-07-15 18:12:24.658542] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.199 [2024-07-15 18:12:24.658558] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.199 [2024-07-15 18:12:24.658565] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.199 [2024-07-15 18:12:24.658741] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.199 [2024-07-15 18:12:24.658918] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.199 [2024-07-15 18:12:24.658926] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.199 [2024-07-15 18:12:24.658932] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.199 [2024-07-15 18:12:24.661758] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.199 [2024-07-15 18:12:24.671266] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.199 [2024-07-15 18:12:24.671699] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.199 [2024-07-15 18:12:24.671715] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.199 [2024-07-15 18:12:24.671724] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.199 [2024-07-15 18:12:24.671901] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.199 [2024-07-15 18:12:24.672078] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.199 [2024-07-15 18:12:24.672086] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.199 [2024-07-15 18:12:24.672092] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.199 [2024-07-15 18:12:24.674916] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.199 [2024-07-15 18:12:24.684425] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.199 [2024-07-15 18:12:24.684898] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.199 [2024-07-15 18:12:24.684914] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.199 [2024-07-15 18:12:24.684921] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.199 [2024-07-15 18:12:24.685097] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.199 [2024-07-15 18:12:24.685279] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.199 [2024-07-15 18:12:24.685288] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.199 [2024-07-15 18:12:24.685294] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.199 [2024-07-15 18:12:24.688117] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.199 [2024-07-15 18:12:24.697243] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.199 [2024-07-15 18:12:24.697729] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.199 [2024-07-15 18:12:24.697770] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.199 [2024-07-15 18:12:24.697791] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.199 [2024-07-15 18:12:24.698341] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.199 [2024-07-15 18:12:24.698513] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.199 [2024-07-15 18:12:24.698521] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.199 [2024-07-15 18:12:24.698527] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.199 [2024-07-15 18:12:24.701196] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.199 [2024-07-15 18:12:24.710149] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.199 [2024-07-15 18:12:24.710608] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.199 [2024-07-15 18:12:24.710623] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.199 [2024-07-15 18:12:24.710630] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.199 [2024-07-15 18:12:24.710800] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.199 [2024-07-15 18:12:24.710971] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.199 [2024-07-15 18:12:24.710982] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.199 [2024-07-15 18:12:24.710988] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.199 [2024-07-15 18:12:24.713662] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.199 [2024-07-15 18:12:24.722928] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.199 [2024-07-15 18:12:24.723404] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.199 [2024-07-15 18:12:24.723445] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.199 [2024-07-15 18:12:24.723467] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.199 [2024-07-15 18:12:24.724043] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.199 [2024-07-15 18:12:24.724621] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.199 [2024-07-15 18:12:24.724629] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.199 [2024-07-15 18:12:24.724635] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.199 [2024-07-15 18:12:24.727311] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.199 [2024-07-15 18:12:24.735760] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.199 [2024-07-15 18:12:24.736252] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.199 [2024-07-15 18:12:24.736269] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.199 [2024-07-15 18:12:24.736276] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.199 [2024-07-15 18:12:24.736449] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.199 [2024-07-15 18:12:24.736621] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.199 [2024-07-15 18:12:24.736629] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.200 [2024-07-15 18:12:24.736635] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.200 [2024-07-15 18:12:24.739334] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.200 [2024-07-15 18:12:24.748622] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.200 [2024-07-15 18:12:24.749103] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.200 [2024-07-15 18:12:24.749160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.200 [2024-07-15 18:12:24.749182] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.200 [2024-07-15 18:12:24.749696] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.200 [2024-07-15 18:12:24.749869] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.200 [2024-07-15 18:12:24.749877] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.200 [2024-07-15 18:12:24.749883] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.200 [2024-07-15 18:12:24.752573] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.200 [2024-07-15 18:12:24.761407] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.200 [2024-07-15 18:12:24.761866] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.200 [2024-07-15 18:12:24.761908] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.200 [2024-07-15 18:12:24.761929] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.200 [2024-07-15 18:12:24.762525] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.200 [2024-07-15 18:12:24.762883] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.200 [2024-07-15 18:12:24.762891] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.200 [2024-07-15 18:12:24.762897] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.200 [2024-07-15 18:12:24.765528] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.200 [2024-07-15 18:12:24.774249] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.200 [2024-07-15 18:12:24.774693] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.200 [2024-07-15 18:12:24.774734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.200 [2024-07-15 18:12:24.774756] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.200 [2024-07-15 18:12:24.775348] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.200 [2024-07-15 18:12:24.775804] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.200 [2024-07-15 18:12:24.775812] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.200 [2024-07-15 18:12:24.775818] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.200 [2024-07-15 18:12:24.778496] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.200 [2024-07-15 18:12:24.787174] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.200 [2024-07-15 18:12:24.787629] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.200 [2024-07-15 18:12:24.787644] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.200 [2024-07-15 18:12:24.787650] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.200 [2024-07-15 18:12:24.787822] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.200 [2024-07-15 18:12:24.787993] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.200 [2024-07-15 18:12:24.788001] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.200 [2024-07-15 18:12:24.788006] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.200 [2024-07-15 18:12:24.790684] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.200 [2024-07-15 18:12:24.800105] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.200 [2024-07-15 18:12:24.800509] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.200 [2024-07-15 18:12:24.800525] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.200 [2024-07-15 18:12:24.800531] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.200 [2024-07-15 18:12:24.800706] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.200 [2024-07-15 18:12:24.800878] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.200 [2024-07-15 18:12:24.800885] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.200 [2024-07-15 18:12:24.800891] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.200 [2024-07-15 18:12:24.803881] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.200 [2024-07-15 18:12:24.812918] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.200 [2024-07-15 18:12:24.813384] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.200 [2024-07-15 18:12:24.813399] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.200 [2024-07-15 18:12:24.813406] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.200 [2024-07-15 18:12:24.813568] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.200 [2024-07-15 18:12:24.813729] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.200 [2024-07-15 18:12:24.813736] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.200 [2024-07-15 18:12:24.813742] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.200 [2024-07-15 18:12:24.816404] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.200 [2024-07-15 18:12:24.825716] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.200 [2024-07-15 18:12:24.826118] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.200 [2024-07-15 18:12:24.826133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.200 [2024-07-15 18:12:24.826140] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.200 [2024-07-15 18:12:24.826317] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.200 [2024-07-15 18:12:24.826488] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.200 [2024-07-15 18:12:24.826496] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.200 [2024-07-15 18:12:24.826501] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.200 [2024-07-15 18:12:24.829171] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.200 [2024-07-15 18:12:24.838593] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.200 [2024-07-15 18:12:24.839020] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.200 [2024-07-15 18:12:24.839036] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.200 [2024-07-15 18:12:24.839042] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.200 [2024-07-15 18:12:24.839204] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.200 [2024-07-15 18:12:24.839393] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.200 [2024-07-15 18:12:24.839402] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.200 [2024-07-15 18:12:24.839411] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.200 [2024-07-15 18:12:24.842076] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.200 [2024-07-15 18:12:24.851495] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.200 [2024-07-15 18:12:24.851971] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.200 [2024-07-15 18:12:24.851987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.200 [2024-07-15 18:12:24.851993] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.200 [2024-07-15 18:12:24.852164] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.200 [2024-07-15 18:12:24.852342] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.200 [2024-07-15 18:12:24.852350] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.200 [2024-07-15 18:12:24.852356] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.200 [2024-07-15 18:12:24.855026] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.200 [2024-07-15 18:12:24.864311] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.200 [2024-07-15 18:12:24.864785] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.200 [2024-07-15 18:12:24.864826] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.200 [2024-07-15 18:12:24.864847] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.200 [2024-07-15 18:12:24.865437] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.200 [2024-07-15 18:12:24.865981] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.200 [2024-07-15 18:12:24.865989] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.200 [2024-07-15 18:12:24.865995] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.200 [2024-07-15 18:12:24.868669] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.200 [2024-07-15 18:12:24.877312] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.200 [2024-07-15 18:12:24.877766] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.200 [2024-07-15 18:12:24.877782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.200 [2024-07-15 18:12:24.877788] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.200 [2024-07-15 18:12:24.877959] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.200 [2024-07-15 18:12:24.878131] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.200 [2024-07-15 18:12:24.878139] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.200 [2024-07-15 18:12:24.878145] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.200 [2024-07-15 18:12:24.880829] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.200 [2024-07-15 18:12:24.890096] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.200 [2024-07-15 18:12:24.890484] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.200 [2024-07-15 18:12:24.890499] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.200 [2024-07-15 18:12:24.890506] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.200 [2024-07-15 18:12:24.890677] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.200 [2024-07-15 18:12:24.890848] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.200 [2024-07-15 18:12:24.890856] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.200 [2024-07-15 18:12:24.890862] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.200 [2024-07-15 18:12:24.893537] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.200 [2024-07-15 18:12:24.903025] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.200 [2024-07-15 18:12:24.903424] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.200 [2024-07-15 18:12:24.903441] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.200 [2024-07-15 18:12:24.903448] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.200 [2024-07-15 18:12:24.903631] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.200 [2024-07-15 18:12:24.903803] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.200 [2024-07-15 18:12:24.903811] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.200 [2024-07-15 18:12:24.903817] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.200 [2024-07-15 18:12:24.906494] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.200 [2024-07-15 18:12:24.915920] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.200 [2024-07-15 18:12:24.916313] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.200 [2024-07-15 18:12:24.916329] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.200 [2024-07-15 18:12:24.916336] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.200 [2024-07-15 18:12:24.916506] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.200 [2024-07-15 18:12:24.916678] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.200 [2024-07-15 18:12:24.916685] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.200 [2024-07-15 18:12:24.916691] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.200 [2024-07-15 18:12:24.919434] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.461 [2024-07-15 18:12:24.928918] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.461 [2024-07-15 18:12:24.929295] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.461 [2024-07-15 18:12:24.929311] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.461 [2024-07-15 18:12:24.929318] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.461 [2024-07-15 18:12:24.929494] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.461 [2024-07-15 18:12:24.929674] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.461 [2024-07-15 18:12:24.929682] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.461 [2024-07-15 18:12:24.929689] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.461 [2024-07-15 18:12:24.932511] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.461 [2024-07-15 18:12:24.941957] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.461 [2024-07-15 18:12:24.942454] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.461 [2024-07-15 18:12:24.942471] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.461 [2024-07-15 18:12:24.942478] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.461 [2024-07-15 18:12:24.942650] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.461 [2024-07-15 18:12:24.942821] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.461 [2024-07-15 18:12:24.942829] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.461 [2024-07-15 18:12:24.942835] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.461 [2024-07-15 18:12:24.945584] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.461 [2024-07-15 18:12:24.954970] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.461 [2024-07-15 18:12:24.955403] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.461 [2024-07-15 18:12:24.955419] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.461 [2024-07-15 18:12:24.955426] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.461 [2024-07-15 18:12:24.955597] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.461 [2024-07-15 18:12:24.955769] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.461 [2024-07-15 18:12:24.955776] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.461 [2024-07-15 18:12:24.955782] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.461 [2024-07-15 18:12:24.958529] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.461 [2024-07-15 18:12:24.967792] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.461 [2024-07-15 18:12:24.968200] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.461 [2024-07-15 18:12:24.968254] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.461 [2024-07-15 18:12:24.968277] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.461 [2024-07-15 18:12:24.968854] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.461 [2024-07-15 18:12:24.969375] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.461 [2024-07-15 18:12:24.969383] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.461 [2024-07-15 18:12:24.969389] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.461 [2024-07-15 18:12:24.973405] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.461 [2024-07-15 18:12:24.981370] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.461 [2024-07-15 18:12:24.981806] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.461 [2024-07-15 18:12:24.981820] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.461 [2024-07-15 18:12:24.981827] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.461 [2024-07-15 18:12:24.981993] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.461 [2024-07-15 18:12:24.982158] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.461 [2024-07-15 18:12:24.982165] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.461 [2024-07-15 18:12:24.982171] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.461 [2024-07-15 18:12:24.984901] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.461 [2024-07-15 18:12:24.994259] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.461 [2024-07-15 18:12:24.994726] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.461 [2024-07-15 18:12:24.994766] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.461 [2024-07-15 18:12:24.994786] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.461 [2024-07-15 18:12:24.995333] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.461 [2024-07-15 18:12:24.995504] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.461 [2024-07-15 18:12:24.995512] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.461 [2024-07-15 18:12:24.995518] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.461 [2024-07-15 18:12:24.998185] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.461 [2024-07-15 18:12:25.007146] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.461 [2024-07-15 18:12:25.007620] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.461 [2024-07-15 18:12:25.007636] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.462 [2024-07-15 18:12:25.007643] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.462 [2024-07-15 18:12:25.007813] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.462 [2024-07-15 18:12:25.007984] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.462 [2024-07-15 18:12:25.007992] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.462 [2024-07-15 18:12:25.007998] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.462 [2024-07-15 18:12:25.010678] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.462 [2024-07-15 18:12:25.019941] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.462 [2024-07-15 18:12:25.020391] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.462 [2024-07-15 18:12:25.020408] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.462 [2024-07-15 18:12:25.020415] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.462 [2024-07-15 18:12:25.020576] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.462 [2024-07-15 18:12:25.020738] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.462 [2024-07-15 18:12:25.020745] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.462 [2024-07-15 18:12:25.020750] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.462 [2024-07-15 18:12:25.023416] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.462 [2024-07-15 18:12:25.032779] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.462 [2024-07-15 18:12:25.033234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.462 [2024-07-15 18:12:25.033249] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.462 [2024-07-15 18:12:25.033271] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.462 [2024-07-15 18:12:25.033442] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.462 [2024-07-15 18:12:25.033614] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.462 [2024-07-15 18:12:25.033622] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.462 [2024-07-15 18:12:25.033628] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.462 [2024-07-15 18:12:25.036324] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.462 [2024-07-15 18:12:25.045566] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.462 [2024-07-15 18:12:25.046013] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.462 [2024-07-15 18:12:25.046028] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.462 [2024-07-15 18:12:25.046035] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.462 [2024-07-15 18:12:25.046206] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.462 [2024-07-15 18:12:25.046384] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.462 [2024-07-15 18:12:25.046392] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.462 [2024-07-15 18:12:25.046398] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.462 [2024-07-15 18:12:25.049069] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.462 [2024-07-15 18:12:25.058483] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.462 [2024-07-15 18:12:25.058943] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.462 [2024-07-15 18:12:25.058958] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.462 [2024-07-15 18:12:25.058965] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.462 [2024-07-15 18:12:25.059136] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.462 [2024-07-15 18:12:25.059317] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.462 [2024-07-15 18:12:25.059326] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.462 [2024-07-15 18:12:25.059332] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.462 [2024-07-15 18:12:25.062001] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.462 [2024-07-15 18:12:25.071263] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.462 [2024-07-15 18:12:25.071637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.462 [2024-07-15 18:12:25.071652] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.462 [2024-07-15 18:12:25.071658] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.462 [2024-07-15 18:12:25.071819] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.462 [2024-07-15 18:12:25.071981] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.462 [2024-07-15 18:12:25.071988] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.462 [2024-07-15 18:12:25.071993] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.462 [2024-07-15 18:12:25.074583] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.462 [2024-07-15 18:12:25.084076] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.462 [2024-07-15 18:12:25.084521] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.462 [2024-07-15 18:12:25.084537] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.462 [2024-07-15 18:12:25.084543] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.462 [2024-07-15 18:12:25.084714] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.462 [2024-07-15 18:12:25.084885] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.462 [2024-07-15 18:12:25.084893] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.462 [2024-07-15 18:12:25.084899] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.462 [2024-07-15 18:12:25.087580] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.462 [2024-07-15 18:12:25.097003] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.462 [2024-07-15 18:12:25.097453] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.462 [2024-07-15 18:12:25.097468] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.462 [2024-07-15 18:12:25.097475] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.462 [2024-07-15 18:12:25.097636] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.462 [2024-07-15 18:12:25.097797] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.462 [2024-07-15 18:12:25.097805] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.462 [2024-07-15 18:12:25.097811] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.462 [2024-07-15 18:12:25.100478] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.462 [2024-07-15 18:12:25.109871] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.462 [2024-07-15 18:12:25.110338] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.462 [2024-07-15 18:12:25.110380] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.462 [2024-07-15 18:12:25.110401] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.462 [2024-07-15 18:12:25.110979] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.462 [2024-07-15 18:12:25.111291] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.462 [2024-07-15 18:12:25.111303] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.462 [2024-07-15 18:12:25.111312] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.462 [2024-07-15 18:12:25.115358] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.462 [2024-07-15 18:12:25.123243] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.462 [2024-07-15 18:12:25.123727] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.462 [2024-07-15 18:12:25.123768] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.462 [2024-07-15 18:12:25.123790] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.462 [2024-07-15 18:12:25.124313] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.462 [2024-07-15 18:12:25.124486] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.462 [2024-07-15 18:12:25.124494] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.462 [2024-07-15 18:12:25.124500] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.462 [2024-07-15 18:12:25.127272] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.462 [2024-07-15 18:12:25.136156] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.462 [2024-07-15 18:12:25.136643] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.462 [2024-07-15 18:12:25.136684] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.462 [2024-07-15 18:12:25.136706] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.462 [2024-07-15 18:12:25.137299] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.462 [2024-07-15 18:12:25.137581] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.462 [2024-07-15 18:12:25.137589] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.462 [2024-07-15 18:12:25.137595] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.462 [2024-07-15 18:12:25.140266] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.462 [2024-07-15 18:12:25.149067] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.462 [2024-07-15 18:12:25.149467] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.463 [2024-07-15 18:12:25.149484] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.463 [2024-07-15 18:12:25.149493] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.463 [2024-07-15 18:12:25.149665] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.463 [2024-07-15 18:12:25.149836] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.463 [2024-07-15 18:12:25.149844] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.463 [2024-07-15 18:12:25.149849] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.463 [2024-07-15 18:12:25.152534] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.463 [2024-07-15 18:12:25.161955] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.463 [2024-07-15 18:12:25.162408] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.463 [2024-07-15 18:12:25.162424] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.463 [2024-07-15 18:12:25.162430] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.463 [2024-07-15 18:12:25.162592] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.463 [2024-07-15 18:12:25.162753] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.463 [2024-07-15 18:12:25.162761] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.463 [2024-07-15 18:12:25.162766] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.463 [2024-07-15 18:12:25.165433] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.463 [2024-07-15 18:12:25.174792] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.463 [2024-07-15 18:12:25.175252] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.463 [2024-07-15 18:12:25.175294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.463 [2024-07-15 18:12:25.175316] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.463 [2024-07-15 18:12:25.175727] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.463 [2024-07-15 18:12:25.175889] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.463 [2024-07-15 18:12:25.175896] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.463 [2024-07-15 18:12:25.175902] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.463 [2024-07-15 18:12:25.178592] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.721 [2024-07-15 18:12:25.187765] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.721 [2024-07-15 18:12:25.188245] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.721 [2024-07-15 18:12:25.188261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.721 [2024-07-15 18:12:25.188268] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.721 [2024-07-15 18:12:25.188444] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.721 [2024-07-15 18:12:25.188622] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.721 [2024-07-15 18:12:25.188633] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.721 [2024-07-15 18:12:25.188639] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.721 [2024-07-15 18:12:25.191468] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.721 [2024-07-15 18:12:25.200724] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.721 [2024-07-15 18:12:25.201132] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.721 [2024-07-15 18:12:25.201148] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.721 [2024-07-15 18:12:25.201155] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.721 [2024-07-15 18:12:25.201335] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.721 [2024-07-15 18:12:25.201511] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.721 [2024-07-15 18:12:25.201519] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.721 [2024-07-15 18:12:25.201525] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.721 [2024-07-15 18:12:25.204313] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.721 [2024-07-15 18:12:25.213815] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.721 [2024-07-15 18:12:25.214295] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.721 [2024-07-15 18:12:25.214338] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.721 [2024-07-15 18:12:25.214359] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.721 [2024-07-15 18:12:25.214855] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.721 [2024-07-15 18:12:25.215027] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.721 [2024-07-15 18:12:25.215035] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.721 [2024-07-15 18:12:25.215041] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.721 [2024-07-15 18:12:25.217778] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.721 [2024-07-15 18:12:25.226618] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.721 [2024-07-15 18:12:25.226992] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.721 [2024-07-15 18:12:25.227007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.721 [2024-07-15 18:12:25.227014] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.721 [2024-07-15 18:12:25.227185] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.721 [2024-07-15 18:12:25.227363] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.721 [2024-07-15 18:12:25.227372] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.721 [2024-07-15 18:12:25.227377] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.721 [2024-07-15 18:12:25.230049] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.721 [2024-07-15 18:12:25.239566] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.721 [2024-07-15 18:12:25.240027] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.721 [2024-07-15 18:12:25.240067] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.721 [2024-07-15 18:12:25.240089] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.721 [2024-07-15 18:12:25.240682] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.721 [2024-07-15 18:12:25.241054] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.721 [2024-07-15 18:12:25.241062] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.721 [2024-07-15 18:12:25.241068] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.721 [2024-07-15 18:12:25.243742] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.721 [2024-07-15 18:12:25.252527] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.721 [2024-07-15 18:12:25.252997] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.721 [2024-07-15 18:12:25.253039] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.721 [2024-07-15 18:12:25.253061] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.721 [2024-07-15 18:12:25.253623] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.721 [2024-07-15 18:12:25.253795] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.722 [2024-07-15 18:12:25.253803] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.722 [2024-07-15 18:12:25.253809] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.722 [2024-07-15 18:12:25.256554] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.722 [2024-07-15 18:12:25.265482] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.722 [2024-07-15 18:12:25.265966] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.722 [2024-07-15 18:12:25.266007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.722 [2024-07-15 18:12:25.266028] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.722 [2024-07-15 18:12:25.266621] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.722 [2024-07-15 18:12:25.266847] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.722 [2024-07-15 18:12:25.266855] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.722 [2024-07-15 18:12:25.266860] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.722 [2024-07-15 18:12:25.269528] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.722 [2024-07-15 18:12:25.278348] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.722 [2024-07-15 18:12:25.278818] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.722 [2024-07-15 18:12:25.278859] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.722 [2024-07-15 18:12:25.278880] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.722 [2024-07-15 18:12:25.279480] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.722 [2024-07-15 18:12:25.280064] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.722 [2024-07-15 18:12:25.280071] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.722 [2024-07-15 18:12:25.280077] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.722 [2024-07-15 18:12:25.282753] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.722 [2024-07-15 18:12:25.291214] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.722 [2024-07-15 18:12:25.291669] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.722 [2024-07-15 18:12:25.291684] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.722 [2024-07-15 18:12:25.291690] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.722 [2024-07-15 18:12:25.291851] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.722 [2024-07-15 18:12:25.292012] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.722 [2024-07-15 18:12:25.292019] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.722 [2024-07-15 18:12:25.292025] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.722 [2024-07-15 18:12:25.294706] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.722 [2024-07-15 18:12:25.304115] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.722 [2024-07-15 18:12:25.304593] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.722 [2024-07-15 18:12:25.304608] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.722 [2024-07-15 18:12:25.304615] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.722 [2024-07-15 18:12:25.304785] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.722 [2024-07-15 18:12:25.304956] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.722 [2024-07-15 18:12:25.304964] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.722 [2024-07-15 18:12:25.304970] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.722 [2024-07-15 18:12:25.307645] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.722 [2024-07-15 18:12:25.316931] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.722 [2024-07-15 18:12:25.317390] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.722 [2024-07-15 18:12:25.317433] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.722 [2024-07-15 18:12:25.317455] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.722 [2024-07-15 18:12:25.318018] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.722 [2024-07-15 18:12:25.318180] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.722 [2024-07-15 18:12:25.318187] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.722 [2024-07-15 18:12:25.318196] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.722 [2024-07-15 18:12:25.320885] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.722 [2024-07-15 18:12:25.329841] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.722 [2024-07-15 18:12:25.330302] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.722 [2024-07-15 18:12:25.330343] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.722 [2024-07-15 18:12:25.330364] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.722 [2024-07-15 18:12:25.330926] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.722 [2024-07-15 18:12:25.331088] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.722 [2024-07-15 18:12:25.331095] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.722 [2024-07-15 18:12:25.331101] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.722 [2024-07-15 18:12:25.333784] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.722 [2024-07-15 18:12:25.342688] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.722 [2024-07-15 18:12:25.343055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.722 [2024-07-15 18:12:25.343070] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.722 [2024-07-15 18:12:25.343076] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.722 [2024-07-15 18:12:25.343254] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.722 [2024-07-15 18:12:25.343435] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.722 [2024-07-15 18:12:25.343443] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.722 [2024-07-15 18:12:25.343449] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.722 [2024-07-15 18:12:25.346119] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.722 [2024-07-15 18:12:25.355540] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.722 [2024-07-15 18:12:25.356030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.722 [2024-07-15 18:12:25.356072] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.722 [2024-07-15 18:12:25.356093] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.722 [2024-07-15 18:12:25.356685] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.722 [2024-07-15 18:12:25.357255] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.722 [2024-07-15 18:12:25.357263] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.722 [2024-07-15 18:12:25.357269] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.722 [2024-07-15 18:12:25.359939] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.722 [2024-07-15 18:12:25.368459] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.722 [2024-07-15 18:12:25.368921] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.722 [2024-07-15 18:12:25.368939] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.722 [2024-07-15 18:12:25.368946] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.722 [2024-07-15 18:12:25.369117] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.722 [2024-07-15 18:12:25.369311] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.722 [2024-07-15 18:12:25.369320] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.722 [2024-07-15 18:12:25.369326] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.722 [2024-07-15 18:12:25.371969] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.722 [2024-07-15 18:12:25.381271] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.722 [2024-07-15 18:12:25.381758] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.722 [2024-07-15 18:12:25.381799] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.722 [2024-07-15 18:12:25.381821] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.722 [2024-07-15 18:12:25.382215] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.722 [2024-07-15 18:12:25.382391] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.722 [2024-07-15 18:12:25.382400] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.722 [2024-07-15 18:12:25.382405] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.722 [2024-07-15 18:12:25.385079] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.722 [2024-07-15 18:12:25.394056] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.722 [2024-07-15 18:12:25.394537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.722 [2024-07-15 18:12:25.394553] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.722 [2024-07-15 18:12:25.394559] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.722 [2024-07-15 18:12:25.394730] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.722 [2024-07-15 18:12:25.394901] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.722 [2024-07-15 18:12:25.394908] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.722 [2024-07-15 18:12:25.394914] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.722 [2024-07-15 18:12:25.397644] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.722 [2024-07-15 18:12:25.406898] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.722 [2024-07-15 18:12:25.407346] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.722 [2024-07-15 18:12:25.407362] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.722 [2024-07-15 18:12:25.407368] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.722 [2024-07-15 18:12:25.407538] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.722 [2024-07-15 18:12:25.407713] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.722 [2024-07-15 18:12:25.407720] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.722 [2024-07-15 18:12:25.407726] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.722 [2024-07-15 18:12:25.410435] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.722 [2024-07-15 18:12:25.419696] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.722 [2024-07-15 18:12:25.420145] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.722 [2024-07-15 18:12:25.420161] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.722 [2024-07-15 18:12:25.420167] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.722 [2024-07-15 18:12:25.420344] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.722 [2024-07-15 18:12:25.420515] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.722 [2024-07-15 18:12:25.420523] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.722 [2024-07-15 18:12:25.420529] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.722 [2024-07-15 18:12:25.423199] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.722 [2024-07-15 18:12:25.432549] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.722 [2024-07-15 18:12:25.433034] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.722 [2024-07-15 18:12:25.433076] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.722 [2024-07-15 18:12:25.433097] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.722 [2024-07-15 18:12:25.433561] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.722 [2024-07-15 18:12:25.433732] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.722 [2024-07-15 18:12:25.433740] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.722 [2024-07-15 18:12:25.433746] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.722 [2024-07-15 18:12:25.436421] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.722 [2024-07-15 18:12:25.445553] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.722 [2024-07-15 18:12:25.446030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.722 [2024-07-15 18:12:25.446045] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.722 [2024-07-15 18:12:25.446052] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.722 [2024-07-15 18:12:25.446235] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.722 [2024-07-15 18:12:25.446414] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.722 [2024-07-15 18:12:25.446421] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.722 [2024-07-15 18:12:25.446428] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.982 [2024-07-15 18:12:25.449259] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.982 [2024-07-15 18:12:25.458526] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.982 [2024-07-15 18:12:25.458961] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.982 [2024-07-15 18:12:25.458977] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.982 [2024-07-15 18:12:25.458984] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.982 [2024-07-15 18:12:25.459160] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.982 [2024-07-15 18:12:25.459342] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.982 [2024-07-15 18:12:25.459351] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.982 [2024-07-15 18:12:25.459357] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.982 [2024-07-15 18:12:25.462106] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.982 [2024-07-15 18:12:25.471481] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.982 [2024-07-15 18:12:25.471950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.982 [2024-07-15 18:12:25.471991] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.982 [2024-07-15 18:12:25.472012] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.982 [2024-07-15 18:12:25.472499] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.982 [2024-07-15 18:12:25.472661] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.982 [2024-07-15 18:12:25.472668] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.982 [2024-07-15 18:12:25.472674] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.982 [2024-07-15 18:12:25.475262] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.982 [2024-07-15 18:12:25.484365] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.982 [2024-07-15 18:12:25.484791] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.982 [2024-07-15 18:12:25.484806] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.983 [2024-07-15 18:12:25.484813] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.983 [2024-07-15 18:12:25.484984] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.983 [2024-07-15 18:12:25.485156] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.983 [2024-07-15 18:12:25.485164] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.983 [2024-07-15 18:12:25.485170] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.983 [2024-07-15 18:12:25.487849] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.983 [2024-07-15 18:12:25.497156] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.983 [2024-07-15 18:12:25.497613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.983 [2024-07-15 18:12:25.497654] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.983 [2024-07-15 18:12:25.497683] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.983 [2024-07-15 18:12:25.498275] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.983 [2024-07-15 18:12:25.498668] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.983 [2024-07-15 18:12:25.498677] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.983 [2024-07-15 18:12:25.498683] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.983 [2024-07-15 18:12:25.501360] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.983 [2024-07-15 18:12:25.510100] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.983 [2024-07-15 18:12:25.510553] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.983 [2024-07-15 18:12:25.510569] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.983 [2024-07-15 18:12:25.510575] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.983 [2024-07-15 18:12:25.510745] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.983 [2024-07-15 18:12:25.510915] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.983 [2024-07-15 18:12:25.510922] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.983 [2024-07-15 18:12:25.510928] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.983 [2024-07-15 18:12:25.513604] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.983 [2024-07-15 18:12:25.523020] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.983 [2024-07-15 18:12:25.523449] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.983 [2024-07-15 18:12:25.523465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.983 [2024-07-15 18:12:25.523471] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.983 [2024-07-15 18:12:25.523641] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.983 [2024-07-15 18:12:25.523811] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.983 [2024-07-15 18:12:25.523819] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.983 [2024-07-15 18:12:25.523825] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.983 [2024-07-15 18:12:25.526495] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.983 [2024-07-15 18:12:25.535925] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.983 [2024-07-15 18:12:25.536388] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.983 [2024-07-15 18:12:25.536432] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.983 [2024-07-15 18:12:25.536455] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.983 [2024-07-15 18:12:25.537032] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.983 [2024-07-15 18:12:25.537626] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.983 [2024-07-15 18:12:25.537637] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.983 [2024-07-15 18:12:25.537643] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.983 [2024-07-15 18:12:25.540318] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.983 [2024-07-15 18:12:25.548768] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.983 [2024-07-15 18:12:25.549228] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.983 [2024-07-15 18:12:25.549244] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.983 [2024-07-15 18:12:25.549251] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.983 [2024-07-15 18:12:25.549422] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.983 [2024-07-15 18:12:25.549592] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.983 [2024-07-15 18:12:25.549600] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.983 [2024-07-15 18:12:25.549606] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.983 [2024-07-15 18:12:25.552277] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.983 [2024-07-15 18:12:25.561704] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.983 [2024-07-15 18:12:25.562074] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.983 [2024-07-15 18:12:25.562089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.983 [2024-07-15 18:12:25.562096] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.983 [2024-07-15 18:12:25.562272] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.983 [2024-07-15 18:12:25.562444] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.983 [2024-07-15 18:12:25.562452] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.983 [2024-07-15 18:12:25.562457] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.983 [2024-07-15 18:12:25.565131] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.983 [2024-07-15 18:12:25.574790] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.983 [2024-07-15 18:12:25.575207] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.983 [2024-07-15 18:12:25.575222] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.983 [2024-07-15 18:12:25.575234] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.983 [2024-07-15 18:12:25.575405] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.983 [2024-07-15 18:12:25.575577] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.983 [2024-07-15 18:12:25.575584] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.983 [2024-07-15 18:12:25.575591] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.983 [2024-07-15 18:12:25.578328] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.983 [2024-07-15 18:12:25.587687] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.983 [2024-07-15 18:12:25.588111] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.983 [2024-07-15 18:12:25.588127] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.983 [2024-07-15 18:12:25.588133] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.983 [2024-07-15 18:12:25.588321] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.983 [2024-07-15 18:12:25.588492] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.983 [2024-07-15 18:12:25.588500] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.983 [2024-07-15 18:12:25.588506] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.983 [2024-07-15 18:12:25.591179] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.983 [2024-07-15 18:12:25.600599] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.983 [2024-07-15 18:12:25.601050] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.983 [2024-07-15 18:12:25.601065] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.983 [2024-07-15 18:12:25.601072] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.983 [2024-07-15 18:12:25.601248] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.983 [2024-07-15 18:12:25.601420] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.983 [2024-07-15 18:12:25.601427] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.983 [2024-07-15 18:12:25.601433] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.983 [2024-07-15 18:12:25.604104] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.983 [2024-07-15 18:12:25.613469] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.983 [2024-07-15 18:12:25.613918] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.983 [2024-07-15 18:12:25.613969] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.983 [2024-07-15 18:12:25.613990] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.983 [2024-07-15 18:12:25.614583] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.983 [2024-07-15 18:12:25.614773] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.983 [2024-07-15 18:12:25.614780] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.983 [2024-07-15 18:12:25.614787] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.983 [2024-07-15 18:12:25.617557] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.983 [2024-07-15 18:12:25.626422] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.983 [2024-07-15 18:12:25.626847] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.983 [2024-07-15 18:12:25.626878] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.984 [2024-07-15 18:12:25.626908] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.984 [2024-07-15 18:12:25.627502] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.984 [2024-07-15 18:12:25.628025] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.984 [2024-07-15 18:12:25.628033] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.984 [2024-07-15 18:12:25.628039] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.984 [2024-07-15 18:12:25.630710] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.984 [2024-07-15 18:12:25.639305] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.984 [2024-07-15 18:12:25.639758] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.984 [2024-07-15 18:12:25.639773] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.984 [2024-07-15 18:12:25.639779] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.984 [2024-07-15 18:12:25.639950] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.984 [2024-07-15 18:12:25.640121] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.984 [2024-07-15 18:12:25.640129] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.984 [2024-07-15 18:12:25.640135] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.984 [2024-07-15 18:12:25.642814] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.984 [2024-07-15 18:12:25.652234] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.984 [2024-07-15 18:12:25.652661] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.984 [2024-07-15 18:12:25.652676] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.984 [2024-07-15 18:12:25.652682] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.984 [2024-07-15 18:12:25.652843] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.984 [2024-07-15 18:12:25.653005] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.984 [2024-07-15 18:12:25.653012] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.984 [2024-07-15 18:12:25.653018] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.984 [2024-07-15 18:12:25.655609] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.984 [2024-07-15 18:12:25.665076] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.984 [2024-07-15 18:12:25.665522] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.984 [2024-07-15 18:12:25.665537] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.984 [2024-07-15 18:12:25.665544] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.984 [2024-07-15 18:12:25.665714] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.984 [2024-07-15 18:12:25.665885] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.984 [2024-07-15 18:12:25.665896] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.984 [2024-07-15 18:12:25.665902] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.984 [2024-07-15 18:12:25.668699] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.984 [2024-07-15 18:12:25.678190] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.984 [2024-07-15 18:12:25.678641] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.984 [2024-07-15 18:12:25.678657] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.984 [2024-07-15 18:12:25.678663] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.984 [2024-07-15 18:12:25.678840] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.984 [2024-07-15 18:12:25.679017] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.984 [2024-07-15 18:12:25.679025] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.984 [2024-07-15 18:12:25.679031] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.984 [2024-07-15 18:12:25.681858] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.984 [2024-07-15 18:12:25.691369] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.984 [2024-07-15 18:12:25.691818] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.984 [2024-07-15 18:12:25.691834] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.984 [2024-07-15 18:12:25.691841] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.984 [2024-07-15 18:12:25.692018] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.984 [2024-07-15 18:12:25.692195] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.984 [2024-07-15 18:12:25.692203] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.984 [2024-07-15 18:12:25.692210] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:31.984 [2024-07-15 18:12:25.695034] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:31.984 [2024-07-15 18:12:25.704544] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:31.984 [2024-07-15 18:12:25.705034] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:31.984 [2024-07-15 18:12:25.705050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:31.984 [2024-07-15 18:12:25.705057] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:31.984 [2024-07-15 18:12:25.705241] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:31.984 [2024-07-15 18:12:25.705419] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:31.984 [2024-07-15 18:12:25.705428] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:31.984 [2024-07-15 18:12:25.705434] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.244 [2024-07-15 18:12:25.708265] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.244 [2024-07-15 18:12:25.717615] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.244 [2024-07-15 18:12:25.718072] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.244 [2024-07-15 18:12:25.718113] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.244 [2024-07-15 18:12:25.718134] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.244 [2024-07-15 18:12:25.718729] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.244 [2024-07-15 18:12:25.719192] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.244 [2024-07-15 18:12:25.719201] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.244 [2024-07-15 18:12:25.719207] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.244 [2024-07-15 18:12:25.722035] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.244 [2024-07-15 18:12:25.730747] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.244 [2024-07-15 18:12:25.731166] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.244 [2024-07-15 18:12:25.731182] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.244 [2024-07-15 18:12:25.731188] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.244 [2024-07-15 18:12:25.731371] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.244 [2024-07-15 18:12:25.731712] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.244 [2024-07-15 18:12:25.731722] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.244 [2024-07-15 18:12:25.731729] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.244 [2024-07-15 18:12:25.734562] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.244 [2024-07-15 18:12:25.743534] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.244 [2024-07-15 18:12:25.743968] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.244 [2024-07-15 18:12:25.743985] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.244 [2024-07-15 18:12:25.743992] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.244 [2024-07-15 18:12:25.744163] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.244 [2024-07-15 18:12:25.744341] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.244 [2024-07-15 18:12:25.744350] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.244 [2024-07-15 18:12:25.744356] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.244 [2024-07-15 18:12:25.747032] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.244 [2024-07-15 18:12:25.756330] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.244 [2024-07-15 18:12:25.756711] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.244 [2024-07-15 18:12:25.756727] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.244 [2024-07-15 18:12:25.756733] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.244 [2024-07-15 18:12:25.756912] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.244 [2024-07-15 18:12:25.757083] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.244 [2024-07-15 18:12:25.757091] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.244 [2024-07-15 18:12:25.757096] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.244 [2024-07-15 18:12:25.759777] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.244 [2024-07-15 18:12:25.769117] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.244 [2024-07-15 18:12:25.769531] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.244 [2024-07-15 18:12:25.769548] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.244 [2024-07-15 18:12:25.769555] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.244 [2024-07-15 18:12:25.769726] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.244 [2024-07-15 18:12:25.769898] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.244 [2024-07-15 18:12:25.769906] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.244 [2024-07-15 18:12:25.769912] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.244 [2024-07-15 18:12:25.772656] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.244 [2024-07-15 18:12:25.782047] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.244 [2024-07-15 18:12:25.782503] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.244 [2024-07-15 18:12:25.782519] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.244 [2024-07-15 18:12:25.782527] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.244 [2024-07-15 18:12:25.782697] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.244 [2024-07-15 18:12:25.782868] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.244 [2024-07-15 18:12:25.782875] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.244 [2024-07-15 18:12:25.782881] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.244 [2024-07-15 18:12:25.785559] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.244 [2024-07-15 18:12:25.795093] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.245 [2024-07-15 18:12:25.795561] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.245 [2024-07-15 18:12:25.795577] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.245 [2024-07-15 18:12:25.795584] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.245 [2024-07-15 18:12:25.795761] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.245 [2024-07-15 18:12:25.795937] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.245 [2024-07-15 18:12:25.795945] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.245 [2024-07-15 18:12:25.795955] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.245 [2024-07-15 18:12:25.798787] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.245 [2024-07-15 18:12:25.808129] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.245 [2024-07-15 18:12:25.808524] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.245 [2024-07-15 18:12:25.808539] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.245 [2024-07-15 18:12:25.808546] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.245 [2024-07-15 18:12:25.808722] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.245 [2024-07-15 18:12:25.808898] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.245 [2024-07-15 18:12:25.808906] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.245 [2024-07-15 18:12:25.808912] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.245 [2024-07-15 18:12:25.811746] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.245 [2024-07-15 18:12:25.821277] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.245 [2024-07-15 18:12:25.821708] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.245 [2024-07-15 18:12:25.821750] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.245 [2024-07-15 18:12:25.821771] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.245 [2024-07-15 18:12:25.822288] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.245 [2024-07-15 18:12:25.822466] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.245 [2024-07-15 18:12:25.822474] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.245 [2024-07-15 18:12:25.822480] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.245 [2024-07-15 18:12:25.825305] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.245 [2024-07-15 18:12:25.834248] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.245 [2024-07-15 18:12:25.834665] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.245 [2024-07-15 18:12:25.834680] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.245 [2024-07-15 18:12:25.834687] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.245 [2024-07-15 18:12:25.834858] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.245 [2024-07-15 18:12:25.835030] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.245 [2024-07-15 18:12:25.835038] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.245 [2024-07-15 18:12:25.835044] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.245 [2024-07-15 18:12:25.837753] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.245 [2024-07-15 18:12:25.847166] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.245 [2024-07-15 18:12:25.847500] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.245 [2024-07-15 18:12:25.847518] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.245 [2024-07-15 18:12:25.847525] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.245 [2024-07-15 18:12:25.847696] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.245 [2024-07-15 18:12:25.847867] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.245 [2024-07-15 18:12:25.847876] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.245 [2024-07-15 18:12:25.847882] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.245 [2024-07-15 18:12:25.850560] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.245 [2024-07-15 18:12:25.859991] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.245 [2024-07-15 18:12:25.860493] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.245 [2024-07-15 18:12:25.860535] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.245 [2024-07-15 18:12:25.860556] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.245 [2024-07-15 18:12:25.861132] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.245 [2024-07-15 18:12:25.861481] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.245 [2024-07-15 18:12:25.861489] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.245 [2024-07-15 18:12:25.861495] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.245 [2024-07-15 18:12:25.864167] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.245 [2024-07-15 18:12:25.872941] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.245 [2024-07-15 18:12:25.873433] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.245 [2024-07-15 18:12:25.873476] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.245 [2024-07-15 18:12:25.873498] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.245 [2024-07-15 18:12:25.874074] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.245 [2024-07-15 18:12:25.874611] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.245 [2024-07-15 18:12:25.874620] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.245 [2024-07-15 18:12:25.874626] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.245 [2024-07-15 18:12:25.877324] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.245 [2024-07-15 18:12:25.885733] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.245 [2024-07-15 18:12:25.886184] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.245 [2024-07-15 18:12:25.886199] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.245 [2024-07-15 18:12:25.886206] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.245 [2024-07-15 18:12:25.886381] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.245 [2024-07-15 18:12:25.886555] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.245 [2024-07-15 18:12:25.886563] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.245 [2024-07-15 18:12:25.886569] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.245 [2024-07-15 18:12:25.889246] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.245 [2024-07-15 18:12:25.898665] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.245 [2024-07-15 18:12:25.899137] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.245 [2024-07-15 18:12:25.899152] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.245 [2024-07-15 18:12:25.899159] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.245 [2024-07-15 18:12:25.899334] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.245 [2024-07-15 18:12:25.899506] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.245 [2024-07-15 18:12:25.899513] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.245 [2024-07-15 18:12:25.899519] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.245 [2024-07-15 18:12:25.902191] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.245 [2024-07-15 18:12:25.911562] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.245 [2024-07-15 18:12:25.912025] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.245 [2024-07-15 18:12:25.912066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.245 [2024-07-15 18:12:25.912088] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.245 [2024-07-15 18:12:25.912536] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.245 [2024-07-15 18:12:25.912708] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.245 [2024-07-15 18:12:25.912716] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.245 [2024-07-15 18:12:25.912722] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.245 [2024-07-15 18:12:25.915396] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.245 [2024-07-15 18:12:25.924371] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.245 [2024-07-15 18:12:25.924771] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.245 [2024-07-15 18:12:25.924786] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.245 [2024-07-15 18:12:25.924793] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.245 [2024-07-15 18:12:25.924963] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.245 [2024-07-15 18:12:25.925135] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.245 [2024-07-15 18:12:25.925143] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.245 [2024-07-15 18:12:25.925149] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.246 [2024-07-15 18:12:25.927894] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.246 [2024-07-15 18:12:25.937163] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.246 [2024-07-15 18:12:25.937664] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.246 [2024-07-15 18:12:25.937680] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.246 [2024-07-15 18:12:25.937686] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.246 [2024-07-15 18:12:25.937857] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.246 [2024-07-15 18:12:25.938027] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.246 [2024-07-15 18:12:25.938035] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.246 [2024-07-15 18:12:25.938041] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.246 [2024-07-15 18:12:25.940719] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.246 [2024-07-15 18:12:25.949955] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.246 [2024-07-15 18:12:25.950456] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.246 [2024-07-15 18:12:25.950474] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.246 [2024-07-15 18:12:25.950480] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.246 [2024-07-15 18:12:25.950651] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.246 [2024-07-15 18:12:25.950823] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.246 [2024-07-15 18:12:25.950831] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.246 [2024-07-15 18:12:25.950836] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.246 [2024-07-15 18:12:25.953609] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.246 [2024-07-15 18:12:25.962850] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.246 [2024-07-15 18:12:25.963332] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.246 [2024-07-15 18:12:25.963349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.246 [2024-07-15 18:12:25.963356] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.246 [2024-07-15 18:12:25.963533] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.246 [2024-07-15 18:12:25.963710] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.246 [2024-07-15 18:12:25.963719] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.246 [2024-07-15 18:12:25.963725] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.246 [2024-07-15 18:12:25.966552] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.504 [2024-07-15 18:12:25.975841] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.504 [2024-07-15 18:12:25.976379] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.504 [2024-07-15 18:12:25.976395] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.504 [2024-07-15 18:12:25.976406] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.504 [2024-07-15 18:12:25.976583] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.504 [2024-07-15 18:12:25.976760] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.504 [2024-07-15 18:12:25.976768] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.504 [2024-07-15 18:12:25.976774] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.504 [2024-07-15 18:12:25.979532] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.504 [2024-07-15 18:12:25.988942] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.504 [2024-07-15 18:12:25.989431] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.504 [2024-07-15 18:12:25.989447] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.504 [2024-07-15 18:12:25.989454] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.504 [2024-07-15 18:12:25.989627] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.504 [2024-07-15 18:12:25.989799] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.504 [2024-07-15 18:12:25.989808] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.504 [2024-07-15 18:12:25.989815] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.504 [2024-07-15 18:12:25.992566] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.504 [2024-07-15 18:12:26.001952] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.504 [2024-07-15 18:12:26.002299] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.504 [2024-07-15 18:12:26.002334] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.504 [2024-07-15 18:12:26.002357] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.504 [2024-07-15 18:12:26.002909] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.504 [2024-07-15 18:12:26.003081] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.504 [2024-07-15 18:12:26.003089] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.504 [2024-07-15 18:12:26.003095] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.504 [2024-07-15 18:12:26.005898] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.504 [2024-07-15 18:12:26.014878] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.504 [2024-07-15 18:12:26.015322] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.504 [2024-07-15 18:12:26.015364] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.504 [2024-07-15 18:12:26.015386] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.504 [2024-07-15 18:12:26.015825] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.504 [2024-07-15 18:12:26.015998] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.504 [2024-07-15 18:12:26.016008] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.504 [2024-07-15 18:12:26.016014] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.504 [2024-07-15 18:12:26.018693] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.504 [2024-07-15 18:12:26.027765] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.504 [2024-07-15 18:12:26.028141] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.504 [2024-07-15 18:12:26.028157] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.504 [2024-07-15 18:12:26.028163] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.504 [2024-07-15 18:12:26.028340] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.504 [2024-07-15 18:12:26.028512] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.504 [2024-07-15 18:12:26.028520] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.504 [2024-07-15 18:12:26.028526] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.504 [2024-07-15 18:12:26.031208] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.504 [2024-07-15 18:12:26.040842] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.504 [2024-07-15 18:12:26.041315] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.504 [2024-07-15 18:12:26.041331] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.504 [2024-07-15 18:12:26.041339] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.504 [2024-07-15 18:12:26.041515] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.504 [2024-07-15 18:12:26.041692] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.504 [2024-07-15 18:12:26.041699] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.504 [2024-07-15 18:12:26.041706] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.504 [2024-07-15 18:12:26.044528] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.504 [2024-07-15 18:12:26.054023] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.504 [2024-07-15 18:12:26.054497] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.505 [2024-07-15 18:12:26.054513] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.505 [2024-07-15 18:12:26.054520] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.505 [2024-07-15 18:12:26.054697] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.505 [2024-07-15 18:12:26.054873] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.505 [2024-07-15 18:12:26.054881] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.505 [2024-07-15 18:12:26.054887] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.505 [2024-07-15 18:12:26.057718] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.505 [2024-07-15 18:12:26.067056] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.505 [2024-07-15 18:12:26.067497] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.505 [2024-07-15 18:12:26.067513] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.505 [2024-07-15 18:12:26.067520] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.505 [2024-07-15 18:12:26.067697] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.505 [2024-07-15 18:12:26.067874] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.505 [2024-07-15 18:12:26.067882] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.505 [2024-07-15 18:12:26.067888] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.505 [2024-07-15 18:12:26.070716] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.505 [2024-07-15 18:12:26.080215] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.505 [2024-07-15 18:12:26.080630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.505 [2024-07-15 18:12:26.080671] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.505 [2024-07-15 18:12:26.080692] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.505 [2024-07-15 18:12:26.081104] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.505 [2024-07-15 18:12:26.081286] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.505 [2024-07-15 18:12:26.081294] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.505 [2024-07-15 18:12:26.081301] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.505 [2024-07-15 18:12:26.084117] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.505 [2024-07-15 18:12:26.093183] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.505 [2024-07-15 18:12:26.093566] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.505 [2024-07-15 18:12:26.093582] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.505 [2024-07-15 18:12:26.093589] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.505 [2024-07-15 18:12:26.093760] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.505 [2024-07-15 18:12:26.093930] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.505 [2024-07-15 18:12:26.093938] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.505 [2024-07-15 18:12:26.093944] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.505 [2024-07-15 18:12:26.096682] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.505 [2024-07-15 18:12:26.106086] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.505 [2024-07-15 18:12:26.106570] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.505 [2024-07-15 18:12:26.106612] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.505 [2024-07-15 18:12:26.106634] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.505 [2024-07-15 18:12:26.107039] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.505 [2024-07-15 18:12:26.107210] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.505 [2024-07-15 18:12:26.107218] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.505 [2024-07-15 18:12:26.107228] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.505 [2024-07-15 18:12:26.109953] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.505 [2024-07-15 18:12:26.118899] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.505 [2024-07-15 18:12:26.119350] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.505 [2024-07-15 18:12:26.119379] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.505 [2024-07-15 18:12:26.119402] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.505 [2024-07-15 18:12:26.119986] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.505 [2024-07-15 18:12:26.120157] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.505 [2024-07-15 18:12:26.120165] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.505 [2024-07-15 18:12:26.120171] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.505 [2024-07-15 18:12:26.122846] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.505 [2024-07-15 18:12:26.131885] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.505 [2024-07-15 18:12:26.132333] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.505 [2024-07-15 18:12:26.132348] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.505 [2024-07-15 18:12:26.132355] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.505 [2024-07-15 18:12:26.132530] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.505 [2024-07-15 18:12:26.132692] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.505 [2024-07-15 18:12:26.132699] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.505 [2024-07-15 18:12:26.132705] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.505 [2024-07-15 18:12:26.135369] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.505 [2024-07-15 18:12:26.144723] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.505 [2024-07-15 18:12:26.145172] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.505 [2024-07-15 18:12:26.145187] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.505 [2024-07-15 18:12:26.145194] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.505 [2024-07-15 18:12:26.145372] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.505 [2024-07-15 18:12:26.145544] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.505 [2024-07-15 18:12:26.145551] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.505 [2024-07-15 18:12:26.145560] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.505 [2024-07-15 18:12:26.148236] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.505 [2024-07-15 18:12:26.157507] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.505 [2024-07-15 18:12:26.157906] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.505 [2024-07-15 18:12:26.157947] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.505 [2024-07-15 18:12:26.157969] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.505 [2024-07-15 18:12:26.158561] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.505 [2024-07-15 18:12:26.159082] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.505 [2024-07-15 18:12:26.159090] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.505 [2024-07-15 18:12:26.159095] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.505 [2024-07-15 18:12:26.161770] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.505 [2024-07-15 18:12:26.170423] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.505 [2024-07-15 18:12:26.170830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.505 [2024-07-15 18:12:26.170845] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.505 [2024-07-15 18:12:26.170851] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.505 [2024-07-15 18:12:26.171013] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.505 [2024-07-15 18:12:26.171175] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.505 [2024-07-15 18:12:26.171182] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.505 [2024-07-15 18:12:26.171188] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.505 [2024-07-15 18:12:26.173884] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.505 [2024-07-15 18:12:26.183246] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.505 [2024-07-15 18:12:26.183701] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.505 [2024-07-15 18:12:26.183716] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.505 [2024-07-15 18:12:26.183723] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.505 [2024-07-15 18:12:26.183894] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.505 [2024-07-15 18:12:26.184065] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.505 [2024-07-15 18:12:26.184073] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.505 [2024-07-15 18:12:26.184079] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.505 [2024-07-15 18:12:26.186758] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.505 [2024-07-15 18:12:26.196020] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.505 [2024-07-15 18:12:26.196478] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.505 [2024-07-15 18:12:26.196528] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.505 [2024-07-15 18:12:26.196549] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.505 [2024-07-15 18:12:26.197099] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.505 [2024-07-15 18:12:26.197277] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.505 [2024-07-15 18:12:26.197285] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.505 [2024-07-15 18:12:26.197291] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.505 [2024-07-15 18:12:26.199957] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.505 [2024-07-15 18:12:26.208971] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.505 [2024-07-15 18:12:26.209469] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.505 [2024-07-15 18:12:26.209512] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.505 [2024-07-15 18:12:26.209533] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.505 [2024-07-15 18:12:26.210023] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.505 [2024-07-15 18:12:26.210195] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.505 [2024-07-15 18:12:26.210202] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.505 [2024-07-15 18:12:26.210209] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.505 [2024-07-15 18:12:26.212888] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.505 [2024-07-15 18:12:26.221793] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.505 [2024-07-15 18:12:26.222281] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.505 [2024-07-15 18:12:26.222329] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.505 [2024-07-15 18:12:26.222350] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.505 [2024-07-15 18:12:26.222927] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.505 [2024-07-15 18:12:26.223471] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.505 [2024-07-15 18:12:26.223482] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.505 [2024-07-15 18:12:26.223491] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.505 [2024-07-15 18:12:26.227546] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.764 [2024-07-15 18:12:26.235324] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.764 [2024-07-15 18:12:26.235801] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.764 [2024-07-15 18:12:26.235843] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.764 [2024-07-15 18:12:26.235866] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.764 [2024-07-15 18:12:26.236367] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.764 [2024-07-15 18:12:26.236538] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.764 [2024-07-15 18:12:26.236546] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.764 [2024-07-15 18:12:26.236552] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.764 [2024-07-15 18:12:26.239290] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.764 [2024-07-15 18:12:26.248120] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.764 [2024-07-15 18:12:26.248566] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.764 [2024-07-15 18:12:26.248582] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.764 [2024-07-15 18:12:26.248588] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.764 [2024-07-15 18:12:26.248760] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.764 [2024-07-15 18:12:26.248931] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.764 [2024-07-15 18:12:26.248938] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.764 [2024-07-15 18:12:26.248945] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.764 [2024-07-15 18:12:26.251626] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.764 [2024-07-15 18:12:26.261155] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.764 [2024-07-15 18:12:26.261605] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.764 [2024-07-15 18:12:26.261621] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.764 [2024-07-15 18:12:26.261628] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.764 [2024-07-15 18:12:26.261804] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.764 [2024-07-15 18:12:26.261981] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.764 [2024-07-15 18:12:26.261989] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.764 [2024-07-15 18:12:26.261995] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.764 [2024-07-15 18:12:26.264725] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.764 [2024-07-15 18:12:26.274108] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.764 [2024-07-15 18:12:26.274562] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.764 [2024-07-15 18:12:26.274610] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.764 [2024-07-15 18:12:26.274632] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.764 [2024-07-15 18:12:26.275175] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.764 [2024-07-15 18:12:26.275353] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.764 [2024-07-15 18:12:26.275361] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.764 [2024-07-15 18:12:26.275370] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.764 [2024-07-15 18:12:26.278040] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.764 [2024-07-15 18:12:26.286943] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.764 [2024-07-15 18:12:26.287390] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.764 [2024-07-15 18:12:26.287444] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.764 [2024-07-15 18:12:26.287466] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.764 [2024-07-15 18:12:26.288044] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.764 [2024-07-15 18:12:26.288543] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.764 [2024-07-15 18:12:26.288552] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.764 [2024-07-15 18:12:26.288558] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.764 [2024-07-15 18:12:26.291234] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.764 [2024-07-15 18:12:26.299732] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.764 [2024-07-15 18:12:26.300166] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.764 [2024-07-15 18:12:26.300181] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.764 [2024-07-15 18:12:26.300188] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.764 [2024-07-15 18:12:26.300368] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.764 [2024-07-15 18:12:26.300539] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.764 [2024-07-15 18:12:26.300547] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.764 [2024-07-15 18:12:26.300553] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.764 [2024-07-15 18:12:26.303228] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.764 [2024-07-15 18:12:26.312662] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.764 [2024-07-15 18:12:26.313122] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.764 [2024-07-15 18:12:26.313165] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.764 [2024-07-15 18:12:26.313186] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.764 [2024-07-15 18:12:26.313613] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.764 [2024-07-15 18:12:26.313785] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.764 [2024-07-15 18:12:26.313792] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.764 [2024-07-15 18:12:26.313798] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.764 [2024-07-15 18:12:26.316467] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.764 [2024-07-15 18:12:26.325549] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.764 [2024-07-15 18:12:26.325952] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.764 [2024-07-15 18:12:26.325969] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.764 [2024-07-15 18:12:26.325976] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.764 [2024-07-15 18:12:26.326138] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.764 [2024-07-15 18:12:26.326306] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.764 [2024-07-15 18:12:26.326314] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.764 [2024-07-15 18:12:26.326320] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.764 [2024-07-15 18:12:26.328901] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.764 [2024-07-15 18:12:26.338459] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.764 [2024-07-15 18:12:26.338928] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.764 [2024-07-15 18:12:26.338970] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.764 [2024-07-15 18:12:26.338991] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.764 [2024-07-15 18:12:26.339519] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.764 [2024-07-15 18:12:26.339681] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.764 [2024-07-15 18:12:26.339689] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.764 [2024-07-15 18:12:26.339694] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.764 [2024-07-15 18:12:26.342282] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.764 [2024-07-15 18:12:26.351299] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.764 [2024-07-15 18:12:26.351730] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.764 [2024-07-15 18:12:26.351759] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.764 [2024-07-15 18:12:26.351782] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.764 [2024-07-15 18:12:26.352360] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.764 [2024-07-15 18:12:26.352523] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.764 [2024-07-15 18:12:26.352531] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.764 [2024-07-15 18:12:26.352537] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.764 [2024-07-15 18:12:26.355123] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.764 [2024-07-15 18:12:26.364088] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.764 [2024-07-15 18:12:26.364524] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.764 [2024-07-15 18:12:26.364540] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.764 [2024-07-15 18:12:26.364547] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.764 [2024-07-15 18:12:26.364718] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.764 [2024-07-15 18:12:26.364893] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.764 [2024-07-15 18:12:26.364901] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.765 [2024-07-15 18:12:26.364907] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.765 [2024-07-15 18:12:26.367586] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.765 [2024-07-15 18:12:26.376997] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.765 [2024-07-15 18:12:26.377412] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.765 [2024-07-15 18:12:26.377428] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.765 [2024-07-15 18:12:26.377435] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.765 [2024-07-15 18:12:26.377606] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.765 [2024-07-15 18:12:26.377777] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.765 [2024-07-15 18:12:26.377785] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.765 [2024-07-15 18:12:26.377790] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.765 [2024-07-15 18:12:26.380527] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.765 [2024-07-15 18:12:26.389798] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.765 [2024-07-15 18:12:26.390261] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.765 [2024-07-15 18:12:26.390304] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.765 [2024-07-15 18:12:26.390325] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.765 [2024-07-15 18:12:26.390599] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.765 [2024-07-15 18:12:26.390760] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.765 [2024-07-15 18:12:26.390768] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.765 [2024-07-15 18:12:26.390773] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.765 [2024-07-15 18:12:26.393440] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.765 [2024-07-15 18:12:26.402734] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.765 [2024-07-15 18:12:26.403222] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.765 [2024-07-15 18:12:26.403275] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.765 [2024-07-15 18:12:26.403296] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.765 [2024-07-15 18:12:26.403791] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.765 [2024-07-15 18:12:26.403962] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.765 [2024-07-15 18:12:26.403970] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.765 [2024-07-15 18:12:26.403976] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.765 [2024-07-15 18:12:26.406648] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.765 [2024-07-15 18:12:26.415565] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.765 [2024-07-15 18:12:26.416062] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.765 [2024-07-15 18:12:26.416078] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.765 [2024-07-15 18:12:26.416085] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.765 [2024-07-15 18:12:26.416263] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.765 [2024-07-15 18:12:26.416434] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.765 [2024-07-15 18:12:26.416443] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.765 [2024-07-15 18:12:26.416449] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.765 [2024-07-15 18:12:26.419121] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.765 [2024-07-15 18:12:26.428410] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.765 [2024-07-15 18:12:26.428794] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.765 [2024-07-15 18:12:26.428809] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.765 [2024-07-15 18:12:26.428816] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.765 [2024-07-15 18:12:26.428977] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.765 [2024-07-15 18:12:26.429138] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.765 [2024-07-15 18:12:26.429146] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.765 [2024-07-15 18:12:26.429151] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.765 [2024-07-15 18:12:26.431873] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.765 [2024-07-15 18:12:26.441404] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.765 [2024-07-15 18:12:26.441889] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.765 [2024-07-15 18:12:26.441930] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.765 [2024-07-15 18:12:26.441951] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.765 [2024-07-15 18:12:26.442536] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.765 [2024-07-15 18:12:26.442768] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.765 [2024-07-15 18:12:26.442776] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.765 [2024-07-15 18:12:26.442782] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.765 [2024-07-15 18:12:26.445456] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.765 [2024-07-15 18:12:26.454259] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.765 [2024-07-15 18:12:26.454729] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.765 [2024-07-15 18:12:26.454744] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.765 [2024-07-15 18:12:26.454754] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.765 [2024-07-15 18:12:26.454926] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.765 [2024-07-15 18:12:26.455099] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.765 [2024-07-15 18:12:26.455107] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.765 [2024-07-15 18:12:26.455112] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.765 [2024-07-15 18:12:26.457847] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.765 [2024-07-15 18:12:26.467213] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.765 [2024-07-15 18:12:26.467688] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.765 [2024-07-15 18:12:26.467731] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.765 [2024-07-15 18:12:26.467752] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.765 [2024-07-15 18:12:26.468345] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.765 [2024-07-15 18:12:26.468930] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.765 [2024-07-15 18:12:26.468938] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.765 [2024-07-15 18:12:26.468944] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.765 [2024-07-15 18:12:26.471618] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:32.765 [2024-07-15 18:12:26.480150] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:32.765 [2024-07-15 18:12:26.480634] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:32.765 [2024-07-15 18:12:26.480650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:32.765 [2024-07-15 18:12:26.480657] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:32.765 [2024-07-15 18:12:26.480833] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:32.765 [2024-07-15 18:12:26.481009] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:32.765 [2024-07-15 18:12:26.481017] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:32.765 [2024-07-15 18:12:26.481023] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:32.765 [2024-07-15 18:12:26.483847] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.023 [2024-07-15 18:12:26.493140] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.023 [2024-07-15 18:12:26.493628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.023 [2024-07-15 18:12:26.493644] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.023 [2024-07-15 18:12:26.493651] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.023 [2024-07-15 18:12:26.493822] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.023 [2024-07-15 18:12:26.493994] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.023 [2024-07-15 18:12:26.494005] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.023 [2024-07-15 18:12:26.494011] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.023 [2024-07-15 18:12:26.496753] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.023 [2024-07-15 18:12:26.505983] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.023 [2024-07-15 18:12:26.506456] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.023 [2024-07-15 18:12:26.506471] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.023 [2024-07-15 18:12:26.506478] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.023 [2024-07-15 18:12:26.506640] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.023 [2024-07-15 18:12:26.506801] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.023 [2024-07-15 18:12:26.506809] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.023 [2024-07-15 18:12:26.506814] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.023 [2024-07-15 18:12:26.509482] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.023 [2024-07-15 18:12:26.518906] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.023 [2024-07-15 18:12:26.519386] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.023 [2024-07-15 18:12:26.519430] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.023 [2024-07-15 18:12:26.519452] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.023 [2024-07-15 18:12:26.519947] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.023 [2024-07-15 18:12:26.520109] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.023 [2024-07-15 18:12:26.520117] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.023 [2024-07-15 18:12:26.520122] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.023 [2024-07-15 18:12:26.522811] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.023 [2024-07-15 18:12:26.531874] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.023 [2024-07-15 18:12:26.532352] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.023 [2024-07-15 18:12:26.532368] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.023 [2024-07-15 18:12:26.532374] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.023 [2024-07-15 18:12:26.532545] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.023 [2024-07-15 18:12:26.532717] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.023 [2024-07-15 18:12:26.532724] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.024 [2024-07-15 18:12:26.532730] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.024 [2024-07-15 18:12:26.535409] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.024 [2024-07-15 18:12:26.544809] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.024 [2024-07-15 18:12:26.545272] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.024 [2024-07-15 18:12:26.545315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.024 [2024-07-15 18:12:26.545337] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.024 [2024-07-15 18:12:26.545913] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.024 [2024-07-15 18:12:26.546151] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.024 [2024-07-15 18:12:26.546159] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.024 [2024-07-15 18:12:26.546164] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.024 [2024-07-15 18:12:26.548858] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.024 [2024-07-15 18:12:26.557694] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.024 [2024-07-15 18:12:26.558150] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.024 [2024-07-15 18:12:26.558166] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.024 [2024-07-15 18:12:26.558173] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.024 [2024-07-15 18:12:26.558349] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.024 [2024-07-15 18:12:26.558520] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.024 [2024-07-15 18:12:26.558530] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.024 [2024-07-15 18:12:26.558537] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.024 [2024-07-15 18:12:26.561204] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.024 [2024-07-15 18:12:26.570628] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.024 [2024-07-15 18:12:26.571082] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.024 [2024-07-15 18:12:26.571097] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.024 [2024-07-15 18:12:26.571104] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.024 [2024-07-15 18:12:26.571280] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.024 [2024-07-15 18:12:26.571452] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.024 [2024-07-15 18:12:26.571459] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.024 [2024-07-15 18:12:26.571465] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.024 [2024-07-15 18:12:26.574153] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.024 [2024-07-15 18:12:26.583520] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.024 [2024-07-15 18:12:26.584001] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.024 [2024-07-15 18:12:26.584043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.024 [2024-07-15 18:12:26.584064] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.024 [2024-07-15 18:12:26.584664] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.024 [2024-07-15 18:12:26.585131] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.024 [2024-07-15 18:12:26.585138] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.024 [2024-07-15 18:12:26.585144] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.024 [2024-07-15 18:12:26.587821] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.024 [2024-07-15 18:12:26.596332] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.024 [2024-07-15 18:12:26.596811] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.024 [2024-07-15 18:12:26.596827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.024 [2024-07-15 18:12:26.596833] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.024 [2024-07-15 18:12:26.597004] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.024 [2024-07-15 18:12:26.597174] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.024 [2024-07-15 18:12:26.597182] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.024 [2024-07-15 18:12:26.597188] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.024 [2024-07-15 18:12:26.599866] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.024 [2024-07-15 18:12:26.609125] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.024 [2024-07-15 18:12:26.609578] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.024 [2024-07-15 18:12:26.609619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.024 [2024-07-15 18:12:26.609641] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.024 [2024-07-15 18:12:26.610202] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.024 [2024-07-15 18:12:26.610380] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.024 [2024-07-15 18:12:26.610388] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.024 [2024-07-15 18:12:26.610394] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.024 [2024-07-15 18:12:26.613063] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.024 [2024-07-15 18:12:26.621979] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.024 [2024-07-15 18:12:26.622463] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.024 [2024-07-15 18:12:26.622506] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.024 [2024-07-15 18:12:26.622527] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.024 [2024-07-15 18:12:26.623107] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.024 [2024-07-15 18:12:26.623499] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.024 [2024-07-15 18:12:26.623508] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.024 [2024-07-15 18:12:26.623516] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.024 [2024-07-15 18:12:26.626185] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.024 [2024-07-15 18:12:26.634766] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.024 [2024-07-15 18:12:26.635170] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.024 [2024-07-15 18:12:26.635184] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.024 [2024-07-15 18:12:26.635190] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.024 [2024-07-15 18:12:26.635380] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.024 [2024-07-15 18:12:26.635551] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.024 [2024-07-15 18:12:26.635559] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.024 [2024-07-15 18:12:26.635565] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.024 [2024-07-15 18:12:26.638236] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.024 [2024-07-15 18:12:26.647675] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.024 [2024-07-15 18:12:26.648141] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.024 [2024-07-15 18:12:26.648182] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.024 [2024-07-15 18:12:26.648203] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.024 [2024-07-15 18:12:26.648804] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.024 [2024-07-15 18:12:26.649057] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.024 [2024-07-15 18:12:26.649067] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.024 [2024-07-15 18:12:26.649077] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.024 [2024-07-15 18:12:26.653120] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.024 [2024-07-15 18:12:26.661161] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.024 [2024-07-15 18:12:26.661650] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.024 [2024-07-15 18:12:26.661691] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.024 [2024-07-15 18:12:26.661713] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.024 [2024-07-15 18:12:26.662130] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.024 [2024-07-15 18:12:26.662308] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.024 [2024-07-15 18:12:26.662316] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.024 [2024-07-15 18:12:26.662322] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.024 [2024-07-15 18:12:26.665031] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.024 [2024-07-15 18:12:26.674159] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.024 [2024-07-15 18:12:26.674650] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.024 [2024-07-15 18:12:26.674692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.024 [2024-07-15 18:12:26.674713] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.024 [2024-07-15 18:12:26.675258] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.024 [2024-07-15 18:12:26.675431] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.024 [2024-07-15 18:12:26.675438] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.024 [2024-07-15 18:12:26.675444] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.024 [2024-07-15 18:12:26.678115] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.024 [2024-07-15 18:12:26.687015] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.024 [2024-07-15 18:12:26.687461] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.024 [2024-07-15 18:12:26.687476] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.024 [2024-07-15 18:12:26.687482] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.024 [2024-07-15 18:12:26.687644] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.024 [2024-07-15 18:12:26.687805] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.024 [2024-07-15 18:12:26.687812] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.024 [2024-07-15 18:12:26.687818] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.024 [2024-07-15 18:12:26.690487] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.024 [2024-07-15 18:12:26.699902] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.024 [2024-07-15 18:12:26.700388] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.024 [2024-07-15 18:12:26.700429] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.024 [2024-07-15 18:12:26.700450] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.024 [2024-07-15 18:12:26.701026] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.024 [2024-07-15 18:12:26.701620] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.024 [2024-07-15 18:12:26.701643] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.024 [2024-07-15 18:12:26.701649] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.024 [2024-07-15 18:12:26.704340] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.024 [2024-07-15 18:12:26.712725] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.024 [2024-07-15 18:12:26.713184] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.024 [2024-07-15 18:12:26.713239] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.024 [2024-07-15 18:12:26.713262] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.024 [2024-07-15 18:12:26.713617] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.024 [2024-07-15 18:12:26.713792] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.024 [2024-07-15 18:12:26.713800] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.024 [2024-07-15 18:12:26.713806] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.024 [2024-07-15 18:12:26.716476] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.024 [2024-07-15 18:12:26.725579] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.024 [2024-07-15 18:12:26.726033] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.024 [2024-07-15 18:12:26.726049] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.024 [2024-07-15 18:12:26.726055] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.024 [2024-07-15 18:12:26.726232] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.024 [2024-07-15 18:12:26.726403] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.024 [2024-07-15 18:12:26.726411] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.024 [2024-07-15 18:12:26.726417] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.024 [2024-07-15 18:12:26.729090] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.024 [2024-07-15 18:12:26.738434] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.024 [2024-07-15 18:12:26.738929] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.024 [2024-07-15 18:12:26.738946] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.024 [2024-07-15 18:12:26.738953] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.024 [2024-07-15 18:12:26.739130] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.024 [2024-07-15 18:12:26.739313] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.024 [2024-07-15 18:12:26.739321] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.024 [2024-07-15 18:12:26.739327] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.024 [2024-07-15 18:12:26.742147] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.283 [2024-07-15 18:12:26.751482] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.283 [2024-07-15 18:12:26.751930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.283 [2024-07-15 18:12:26.751987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.283 [2024-07-15 18:12:26.752009] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.283 [2024-07-15 18:12:26.752534] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.283 [2024-07-15 18:12:26.752706] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.283 [2024-07-15 18:12:26.752714] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.283 [2024-07-15 18:12:26.752721] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.283 [2024-07-15 18:12:26.755472] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.283 [2024-07-15 18:12:26.764342] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.283 [2024-07-15 18:12:26.764816] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.283 [2024-07-15 18:12:26.764857] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.283 [2024-07-15 18:12:26.764878] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.283 [2024-07-15 18:12:26.765410] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.283 [2024-07-15 18:12:26.765582] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.283 [2024-07-15 18:12:26.765590] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.283 [2024-07-15 18:12:26.765596] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.283 [2024-07-15 18:12:26.768292] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.283 [2024-07-15 18:12:26.777255] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.283 [2024-07-15 18:12:26.777737] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.283 [2024-07-15 18:12:26.777779] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.283 [2024-07-15 18:12:26.777801] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.283 [2024-07-15 18:12:26.778393] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.283 [2024-07-15 18:12:26.778628] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.283 [2024-07-15 18:12:26.778636] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.283 [2024-07-15 18:12:26.778642] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.283 [2024-07-15 18:12:26.781335] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.283 [2024-07-15 18:12:26.790128] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.283 [2024-07-15 18:12:26.790580] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.283 [2024-07-15 18:12:26.790596] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.283 [2024-07-15 18:12:26.790603] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.283 [2024-07-15 18:12:26.790774] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.283 [2024-07-15 18:12:26.790945] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.283 [2024-07-15 18:12:26.790952] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.283 [2024-07-15 18:12:26.790958] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.283 [2024-07-15 18:12:26.793640] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.283 [2024-07-15 18:12:26.802975] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.284 [2024-07-15 18:12:26.803382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.284 [2024-07-15 18:12:26.803401] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.284 [2024-07-15 18:12:26.803407] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.284 [2024-07-15 18:12:26.803569] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.284 [2024-07-15 18:12:26.803732] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.284 [2024-07-15 18:12:26.803739] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.284 [2024-07-15 18:12:26.803745] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.284 [2024-07-15 18:12:26.806415] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.284 [2024-07-15 18:12:26.815798] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.284 [2024-07-15 18:12:26.816276] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.284 [2024-07-15 18:12:26.816291] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.284 [2024-07-15 18:12:26.816298] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.284 [2024-07-15 18:12:26.816476] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.284 [2024-07-15 18:12:26.816638] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.284 [2024-07-15 18:12:26.816645] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.284 [2024-07-15 18:12:26.816651] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.284 [2024-07-15 18:12:26.819316] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.284 [2024-07-15 18:12:26.828601] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.284 [2024-07-15 18:12:26.829056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.284 [2024-07-15 18:12:26.829098] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.284 [2024-07-15 18:12:26.829120] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.284 [2024-07-15 18:12:26.829569] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.284 [2024-07-15 18:12:26.829741] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.284 [2024-07-15 18:12:26.829748] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.284 [2024-07-15 18:12:26.829754] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.284 [2024-07-15 18:12:26.832449] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.284 [2024-07-15 18:12:26.841474] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.284 [2024-07-15 18:12:26.841804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.284 [2024-07-15 18:12:26.841820] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.284 [2024-07-15 18:12:26.841826] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.284 [2024-07-15 18:12:26.841988] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.284 [2024-07-15 18:12:26.842153] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.284 [2024-07-15 18:12:26.842161] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.284 [2024-07-15 18:12:26.842166] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.284 [2024-07-15 18:12:26.844858] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.284 [2024-07-15 18:12:26.854294] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.284 [2024-07-15 18:12:26.854771] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.284 [2024-07-15 18:12:26.854786] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.284 [2024-07-15 18:12:26.854793] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.284 [2024-07-15 18:12:26.854964] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.284 [2024-07-15 18:12:26.855136] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.284 [2024-07-15 18:12:26.855143] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.284 [2024-07-15 18:12:26.855149] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.284 [2024-07-15 18:12:26.857835] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.284 [2024-07-15 18:12:26.867171] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.284 [2024-07-15 18:12:26.867599] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.284 [2024-07-15 18:12:26.867632] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.284 [2024-07-15 18:12:26.867654] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.284 [2024-07-15 18:12:26.868244] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.284 [2024-07-15 18:12:26.868826] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.284 [2024-07-15 18:12:26.868850] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.284 [2024-07-15 18:12:26.868870] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.284 [2024-07-15 18:12:26.871559] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.284 [2024-07-15 18:12:26.880038] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.284 [2024-07-15 18:12:26.880513] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.284 [2024-07-15 18:12:26.880549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.284 [2024-07-15 18:12:26.880571] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.284 [2024-07-15 18:12:26.881149] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.284 [2024-07-15 18:12:26.881743] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.284 [2024-07-15 18:12:26.881768] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.284 [2024-07-15 18:12:26.881788] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.284 [2024-07-15 18:12:26.884511] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.284 [2024-07-15 18:12:26.892860] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.284 [2024-07-15 18:12:26.893335] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.284 [2024-07-15 18:12:26.893378] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.284 [2024-07-15 18:12:26.893400] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.284 [2024-07-15 18:12:26.893980] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.284 [2024-07-15 18:12:26.894188] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.284 [2024-07-15 18:12:26.894195] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.284 [2024-07-15 18:12:26.894201] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.284 [2024-07-15 18:12:26.896897] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.284 [2024-07-15 18:12:26.905697] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.284 [2024-07-15 18:12:26.906168] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.284 [2024-07-15 18:12:26.906209] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.284 [2024-07-15 18:12:26.906244] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.284 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/bdevperf.sh: line 35: 739498 Killed "${NVMF_APP[@]}" "$@" 00:26:33.284 [2024-07-15 18:12:26.906781] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.284 [2024-07-15 18:12:26.906953] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.284 [2024-07-15 18:12:26.906961] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.284 [2024-07-15 18:12:26.906967] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.284 18:12:26 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@36 -- # tgt_init 00:26:33.284 18:12:26 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@15 -- # nvmfappstart -m 0xE 00:26:33.284 18:12:26 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:26:33.284 18:12:26 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@722 -- # xtrace_disable 00:26:33.284 18:12:26 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:26:33.284 [2024-07-15 18:12:26.909707] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.284 18:12:26 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@481 -- # nvmfpid=740910 00:26:33.284 18:12:26 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@482 -- # waitforlisten 740910 00:26:33.284 18:12:26 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xE 00:26:33.284 18:12:26 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@829 -- # '[' -z 740910 ']' 00:26:33.284 18:12:26 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:33.284 18:12:26 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@834 -- # local max_retries=100 00:26:33.284 18:12:26 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:33.284 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:33.284 18:12:26 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@838 -- # xtrace_disable 00:26:33.284 18:12:26 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:26:33.284 [2024-07-15 18:12:26.918885] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.284 [2024-07-15 18:12:26.919373] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.284 [2024-07-15 18:12:26.919390] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.284 [2024-07-15 18:12:26.919397] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.285 [2024-07-15 18:12:26.919575] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.285 [2024-07-15 18:12:26.919752] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.285 [2024-07-15 18:12:26.919760] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.285 [2024-07-15 18:12:26.919766] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.285 [2024-07-15 18:12:26.922590] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.285 [2024-07-15 18:12:26.931926] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.285 [2024-07-15 18:12:26.932370] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.285 [2024-07-15 18:12:26.932385] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.285 [2024-07-15 18:12:26.932392] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.285 [2024-07-15 18:12:26.932568] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.285 [2024-07-15 18:12:26.932744] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.285 [2024-07-15 18:12:26.932752] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.285 [2024-07-15 18:12:26.932758] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.285 [2024-07-15 18:12:26.935585] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.285 [2024-07-15 18:12:26.945092] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.285 [2024-07-15 18:12:26.945570] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.285 [2024-07-15 18:12:26.945586] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.285 [2024-07-15 18:12:26.945593] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.285 [2024-07-15 18:12:26.945769] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.285 [2024-07-15 18:12:26.945945] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.285 [2024-07-15 18:12:26.945953] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.285 [2024-07-15 18:12:26.945959] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.285 [2024-07-15 18:12:26.948791] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.285 [2024-07-15 18:12:26.958068] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.285 [2024-07-15 18:12:26.958559] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.285 [2024-07-15 18:12:26.958575] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.285 [2024-07-15 18:12:26.958582] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.285 [2024-07-15 18:12:26.958762] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.285 [2024-07-15 18:12:26.958939] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.285 [2024-07-15 18:12:26.958947] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.285 [2024-07-15 18:12:26.958952] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.285 [2024-07-15 18:12:26.961810] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.285 [2024-07-15 18:12:26.961843] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:26:33.285 [2024-07-15 18:12:26.961899] [ DPDK EAL parameters: nvmf -c 0xE --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:33.285 [2024-07-15 18:12:26.971126] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.285 [2024-07-15 18:12:26.971561] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.285 [2024-07-15 18:12:26.971577] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.285 [2024-07-15 18:12:26.971584] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.285 [2024-07-15 18:12:26.971761] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.285 [2024-07-15 18:12:26.971937] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.285 [2024-07-15 18:12:26.971945] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.285 [2024-07-15 18:12:26.971951] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.285 [2024-07-15 18:12:26.974753] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.285 [2024-07-15 18:12:26.984286] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.285 [2024-07-15 18:12:26.984771] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.285 [2024-07-15 18:12:26.984788] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.285 [2024-07-15 18:12:26.984795] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.285 [2024-07-15 18:12:26.984971] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.285 [2024-07-15 18:12:26.985149] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.285 [2024-07-15 18:12:26.985157] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.285 [2024-07-15 18:12:26.985163] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.285 [2024-07-15 18:12:26.987980] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.285 EAL: No free 2048 kB hugepages reported on node 1 00:26:33.285 [2024-07-15 18:12:26.997332] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.285 [2024-07-15 18:12:26.997776] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.285 [2024-07-15 18:12:26.997792] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.285 [2024-07-15 18:12:26.997799] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.285 [2024-07-15 18:12:26.997979] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.285 [2024-07-15 18:12:26.998155] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.285 [2024-07-15 18:12:26.998163] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.285 [2024-07-15 18:12:26.998169] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.285 [2024-07-15 18:12:27.000997] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.544 [2024-07-15 18:12:27.010490] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.544 [2024-07-15 18:12:27.010951] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.544 [2024-07-15 18:12:27.010967] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.544 [2024-07-15 18:12:27.010974] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.544 [2024-07-15 18:12:27.011150] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.544 [2024-07-15 18:12:27.011329] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.544 [2024-07-15 18:12:27.011338] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.544 [2024-07-15 18:12:27.011344] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.544 [2024-07-15 18:12:27.014165] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.544 [2024-07-15 18:12:27.018544] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 3 00:26:33.544 [2024-07-15 18:12:27.023591] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.544 [2024-07-15 18:12:27.024078] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.544 [2024-07-15 18:12:27.024093] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.544 [2024-07-15 18:12:27.024101] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.544 [2024-07-15 18:12:27.024283] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.544 [2024-07-15 18:12:27.024460] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.544 [2024-07-15 18:12:27.024468] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.544 [2024-07-15 18:12:27.024474] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.545 [2024-07-15 18:12:27.027265] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.545 [2024-07-15 18:12:27.036617] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.545 [2024-07-15 18:12:27.037049] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.545 [2024-07-15 18:12:27.037065] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.545 [2024-07-15 18:12:27.037071] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.545 [2024-07-15 18:12:27.037248] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.545 [2024-07-15 18:12:27.037419] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.545 [2024-07-15 18:12:27.037431] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.545 [2024-07-15 18:12:27.037437] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.545 [2024-07-15 18:12:27.040264] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.545 [2024-07-15 18:12:27.049693] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.545 [2024-07-15 18:12:27.050146] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.545 [2024-07-15 18:12:27.050161] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.545 [2024-07-15 18:12:27.050168] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.545 [2024-07-15 18:12:27.050350] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.545 [2024-07-15 18:12:27.050527] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.545 [2024-07-15 18:12:27.050535] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.545 [2024-07-15 18:12:27.050542] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.545 [2024-07-15 18:12:27.053332] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.545 [2024-07-15 18:12:27.062792] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.545 [2024-07-15 18:12:27.063236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.545 [2024-07-15 18:12:27.063256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.545 [2024-07-15 18:12:27.063265] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.545 [2024-07-15 18:12:27.063445] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.545 [2024-07-15 18:12:27.063623] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.545 [2024-07-15 18:12:27.063631] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.545 [2024-07-15 18:12:27.063639] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.545 [2024-07-15 18:12:27.066427] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.545 [2024-07-15 18:12:27.075947] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.545 [2024-07-15 18:12:27.076328] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.545 [2024-07-15 18:12:27.076345] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.545 [2024-07-15 18:12:27.076352] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.545 [2024-07-15 18:12:27.076529] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.545 [2024-07-15 18:12:27.076706] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.545 [2024-07-15 18:12:27.076714] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.545 [2024-07-15 18:12:27.076720] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.545 [2024-07-15 18:12:27.079549] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.545 [2024-07-15 18:12:27.089055] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.545 [2024-07-15 18:12:27.089534] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.545 [2024-07-15 18:12:27.089550] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.545 [2024-07-15 18:12:27.089557] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.545 [2024-07-15 18:12:27.089733] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.545 [2024-07-15 18:12:27.089911] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.545 [2024-07-15 18:12:27.089919] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.545 [2024-07-15 18:12:27.089925] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.545 [2024-07-15 18:12:27.092751] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.545 [2024-07-15 18:12:27.099839] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:26:33.545 [2024-07-15 18:12:27.099864] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:26:33.545 [2024-07-15 18:12:27.099871] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:26:33.545 [2024-07-15 18:12:27.099878] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:26:33.545 [2024-07-15 18:12:27.099883] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:26:33.545 [2024-07-15 18:12:27.099922] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:26:33.545 [2024-07-15 18:12:27.100005] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:26:33.545 [2024-07-15 18:12:27.100007] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:26:33.545 [2024-07-15 18:12:27.102266] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.545 [2024-07-15 18:12:27.102608] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.545 [2024-07-15 18:12:27.102624] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.545 [2024-07-15 18:12:27.102631] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.545 [2024-07-15 18:12:27.102808] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.545 [2024-07-15 18:12:27.102985] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.545 [2024-07-15 18:12:27.102993] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.545 [2024-07-15 18:12:27.103000] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.545 [2024-07-15 18:12:27.105826] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.545 [2024-07-15 18:12:27.115330] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.545 [2024-07-15 18:12:27.115806] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.545 [2024-07-15 18:12:27.115825] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.545 [2024-07-15 18:12:27.115832] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.545 [2024-07-15 18:12:27.116009] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.545 [2024-07-15 18:12:27.116185] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.545 [2024-07-15 18:12:27.116199] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.545 [2024-07-15 18:12:27.116206] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.545 [2024-07-15 18:12:27.119051] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.545 [2024-07-15 18:12:27.128397] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.545 [2024-07-15 18:12:27.128813] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.545 [2024-07-15 18:12:27.128831] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.545 [2024-07-15 18:12:27.128840] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.545 [2024-07-15 18:12:27.129019] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.545 [2024-07-15 18:12:27.129196] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.545 [2024-07-15 18:12:27.129204] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.545 [2024-07-15 18:12:27.129212] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.545 [2024-07-15 18:12:27.132033] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.545 [2024-07-15 18:12:27.141545] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.545 [2024-07-15 18:12:27.141920] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.545 [2024-07-15 18:12:27.141937] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.545 [2024-07-15 18:12:27.141944] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.545 [2024-07-15 18:12:27.142122] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.545 [2024-07-15 18:12:27.142304] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.545 [2024-07-15 18:12:27.142313] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.545 [2024-07-15 18:12:27.142320] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.545 [2024-07-15 18:12:27.145148] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.545 [2024-07-15 18:12:27.154690] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.545 [2024-07-15 18:12:27.155150] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.545 [2024-07-15 18:12:27.155168] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.545 [2024-07-15 18:12:27.155175] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.545 [2024-07-15 18:12:27.155365] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.545 [2024-07-15 18:12:27.155543] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.545 [2024-07-15 18:12:27.155551] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.545 [2024-07-15 18:12:27.155559] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.545 [2024-07-15 18:12:27.158383] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.545 [2024-07-15 18:12:27.167887] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.545 [2024-07-15 18:12:27.168293] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.545 [2024-07-15 18:12:27.168309] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.545 [2024-07-15 18:12:27.168316] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.545 [2024-07-15 18:12:27.168494] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.545 [2024-07-15 18:12:27.168671] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.545 [2024-07-15 18:12:27.168679] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.545 [2024-07-15 18:12:27.168685] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.545 [2024-07-15 18:12:27.171508] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.545 [2024-07-15 18:12:27.181014] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.545 [2024-07-15 18:12:27.181430] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.545 [2024-07-15 18:12:27.181447] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.545 [2024-07-15 18:12:27.181454] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.545 [2024-07-15 18:12:27.181631] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.545 [2024-07-15 18:12:27.181808] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.545 [2024-07-15 18:12:27.181816] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.545 [2024-07-15 18:12:27.181822] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.545 [2024-07-15 18:12:27.184644] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.545 [2024-07-15 18:12:27.194160] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.545 [2024-07-15 18:12:27.194564] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.545 [2024-07-15 18:12:27.194580] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.545 [2024-07-15 18:12:27.194588] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.545 [2024-07-15 18:12:27.194764] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.545 [2024-07-15 18:12:27.194941] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.545 [2024-07-15 18:12:27.194950] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.545 [2024-07-15 18:12:27.194956] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.545 [2024-07-15 18:12:27.197785] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.545 [2024-07-15 18:12:27.207296] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.545 [2024-07-15 18:12:27.207669] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.545 [2024-07-15 18:12:27.207685] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.545 [2024-07-15 18:12:27.207692] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.545 [2024-07-15 18:12:27.207872] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.545 [2024-07-15 18:12:27.208048] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.545 [2024-07-15 18:12:27.208056] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.545 [2024-07-15 18:12:27.208063] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.545 [2024-07-15 18:12:27.210886] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.545 [2024-07-15 18:12:27.220384] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.545 [2024-07-15 18:12:27.220702] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.545 [2024-07-15 18:12:27.220718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.545 [2024-07-15 18:12:27.220725] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.545 [2024-07-15 18:12:27.220901] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.545 [2024-07-15 18:12:27.221078] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.545 [2024-07-15 18:12:27.221086] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.545 [2024-07-15 18:12:27.221092] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.545 [2024-07-15 18:12:27.223915] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.545 [2024-07-15 18:12:27.233415] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.545 [2024-07-15 18:12:27.233789] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.545 [2024-07-15 18:12:27.233804] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.545 [2024-07-15 18:12:27.233811] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.545 [2024-07-15 18:12:27.233987] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.545 [2024-07-15 18:12:27.234162] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.545 [2024-07-15 18:12:27.234170] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.545 [2024-07-15 18:12:27.234176] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.545 [2024-07-15 18:12:27.236999] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.546 [2024-07-15 18:12:27.246509] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.546 [2024-07-15 18:12:27.246878] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.546 [2024-07-15 18:12:27.246894] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.546 [2024-07-15 18:12:27.246901] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.546 [2024-07-15 18:12:27.247077] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.546 [2024-07-15 18:12:27.247257] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.546 [2024-07-15 18:12:27.247265] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.546 [2024-07-15 18:12:27.247275] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.546 [2024-07-15 18:12:27.250099] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.546 [2024-07-15 18:12:27.259620] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.546 [2024-07-15 18:12:27.259936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.546 [2024-07-15 18:12:27.259952] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.546 [2024-07-15 18:12:27.259959] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.546 [2024-07-15 18:12:27.260135] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.546 [2024-07-15 18:12:27.260318] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.546 [2024-07-15 18:12:27.260326] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.546 [2024-07-15 18:12:27.260332] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.546 [2024-07-15 18:12:27.263151] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.804 [2024-07-15 18:12:27.272667] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.804 [2024-07-15 18:12:27.272977] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.804 [2024-07-15 18:12:27.272993] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.804 [2024-07-15 18:12:27.273000] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.804 [2024-07-15 18:12:27.273177] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.804 [2024-07-15 18:12:27.273364] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.804 [2024-07-15 18:12:27.273373] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.804 [2024-07-15 18:12:27.273380] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.804 [2024-07-15 18:12:27.276197] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.805 [2024-07-15 18:12:27.285702] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.805 [2024-07-15 18:12:27.286072] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.805 [2024-07-15 18:12:27.286088] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.805 [2024-07-15 18:12:27.286095] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.805 [2024-07-15 18:12:27.286276] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.805 [2024-07-15 18:12:27.286453] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.805 [2024-07-15 18:12:27.286461] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.805 [2024-07-15 18:12:27.286467] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.805 [2024-07-15 18:12:27.289292] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.805 [2024-07-15 18:12:27.298831] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.805 [2024-07-15 18:12:27.299204] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.805 [2024-07-15 18:12:27.299228] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.805 [2024-07-15 18:12:27.299236] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.805 [2024-07-15 18:12:27.299412] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.805 [2024-07-15 18:12:27.299589] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.805 [2024-07-15 18:12:27.299597] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.805 [2024-07-15 18:12:27.299603] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.805 [2024-07-15 18:12:27.302431] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.805 [2024-07-15 18:12:27.311931] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.805 [2024-07-15 18:12:27.312348] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.805 [2024-07-15 18:12:27.312365] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.805 [2024-07-15 18:12:27.312372] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.805 [2024-07-15 18:12:27.312548] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.805 [2024-07-15 18:12:27.312725] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.805 [2024-07-15 18:12:27.312733] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.805 [2024-07-15 18:12:27.312739] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.805 [2024-07-15 18:12:27.315575] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.805 [2024-07-15 18:12:27.325103] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.805 [2024-07-15 18:12:27.325475] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.805 [2024-07-15 18:12:27.325492] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.805 [2024-07-15 18:12:27.325499] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.805 [2024-07-15 18:12:27.325674] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.805 [2024-07-15 18:12:27.325851] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.805 [2024-07-15 18:12:27.325860] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.805 [2024-07-15 18:12:27.325866] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.805 [2024-07-15 18:12:27.328692] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.805 [2024-07-15 18:12:27.338196] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.805 [2024-07-15 18:12:27.338667] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.805 [2024-07-15 18:12:27.338683] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.805 [2024-07-15 18:12:27.338690] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.805 [2024-07-15 18:12:27.338866] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.805 [2024-07-15 18:12:27.339046] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.805 [2024-07-15 18:12:27.339054] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.805 [2024-07-15 18:12:27.339061] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.805 [2024-07-15 18:12:27.341885] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.805 [2024-07-15 18:12:27.351229] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.805 [2024-07-15 18:12:27.351622] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.805 [2024-07-15 18:12:27.351638] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.805 [2024-07-15 18:12:27.351645] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.805 [2024-07-15 18:12:27.351822] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.805 [2024-07-15 18:12:27.351998] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.805 [2024-07-15 18:12:27.352006] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.805 [2024-07-15 18:12:27.352012] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.805 [2024-07-15 18:12:27.354840] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.805 [2024-07-15 18:12:27.364363] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.805 [2024-07-15 18:12:27.364787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.805 [2024-07-15 18:12:27.364803] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.805 [2024-07-15 18:12:27.364809] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.805 [2024-07-15 18:12:27.364987] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.805 [2024-07-15 18:12:27.365163] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.805 [2024-07-15 18:12:27.365171] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.805 [2024-07-15 18:12:27.365176] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.805 [2024-07-15 18:12:27.367997] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.805 [2024-07-15 18:12:27.377504] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.805 [2024-07-15 18:12:27.377819] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.805 [2024-07-15 18:12:27.377835] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.805 [2024-07-15 18:12:27.377842] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.805 [2024-07-15 18:12:27.378018] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.805 [2024-07-15 18:12:27.378195] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.805 [2024-07-15 18:12:27.378203] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.805 [2024-07-15 18:12:27.378210] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.805 [2024-07-15 18:12:27.381034] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.805 [2024-07-15 18:12:27.390541] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.805 [2024-07-15 18:12:27.390866] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.805 [2024-07-15 18:12:27.390882] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.805 [2024-07-15 18:12:27.390889] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.805 [2024-07-15 18:12:27.391064] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.805 [2024-07-15 18:12:27.391247] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.805 [2024-07-15 18:12:27.391255] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.805 [2024-07-15 18:12:27.391261] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.805 [2024-07-15 18:12:27.394081] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.805 [2024-07-15 18:12:27.403574] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.805 [2024-07-15 18:12:27.403870] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.805 [2024-07-15 18:12:27.403886] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.805 [2024-07-15 18:12:27.403893] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.805 [2024-07-15 18:12:27.404069] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.805 [2024-07-15 18:12:27.404251] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.805 [2024-07-15 18:12:27.404260] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.805 [2024-07-15 18:12:27.404266] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.805 [2024-07-15 18:12:27.407087] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.805 [2024-07-15 18:12:27.416765] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.805 [2024-07-15 18:12:27.417180] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.805 [2024-07-15 18:12:27.417196] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.805 [2024-07-15 18:12:27.417202] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.805 [2024-07-15 18:12:27.417385] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.805 [2024-07-15 18:12:27.417563] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.805 [2024-07-15 18:12:27.417570] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.805 [2024-07-15 18:12:27.417577] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.805 [2024-07-15 18:12:27.420402] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.805 [2024-07-15 18:12:27.429898] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.805 [2024-07-15 18:12:27.430228] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.805 [2024-07-15 18:12:27.430244] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.805 [2024-07-15 18:12:27.430254] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.805 [2024-07-15 18:12:27.430431] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.805 [2024-07-15 18:12:27.430611] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.805 [2024-07-15 18:12:27.430618] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.805 [2024-07-15 18:12:27.430624] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.805 [2024-07-15 18:12:27.433451] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.805 [2024-07-15 18:12:27.442949] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.805 [2024-07-15 18:12:27.443325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.805 [2024-07-15 18:12:27.443342] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.805 [2024-07-15 18:12:27.443348] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.805 [2024-07-15 18:12:27.443524] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.805 [2024-07-15 18:12:27.443701] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.805 [2024-07-15 18:12:27.443709] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.805 [2024-07-15 18:12:27.443715] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.805 [2024-07-15 18:12:27.446541] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.805 [2024-07-15 18:12:27.456055] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.805 [2024-07-15 18:12:27.456382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.805 [2024-07-15 18:12:27.456398] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.805 [2024-07-15 18:12:27.456405] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.805 [2024-07-15 18:12:27.456581] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.805 [2024-07-15 18:12:27.456762] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.805 [2024-07-15 18:12:27.456770] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.805 [2024-07-15 18:12:27.456776] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.805 [2024-07-15 18:12:27.459604] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.805 [2024-07-15 18:12:27.469104] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.805 [2024-07-15 18:12:27.469387] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.805 [2024-07-15 18:12:27.469403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.805 [2024-07-15 18:12:27.469410] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.805 [2024-07-15 18:12:27.469585] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.805 [2024-07-15 18:12:27.469761] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.805 [2024-07-15 18:12:27.469772] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.805 [2024-07-15 18:12:27.469778] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.805 [2024-07-15 18:12:27.472601] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.805 [2024-07-15 18:12:27.482276] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.805 [2024-07-15 18:12:27.482625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.805 [2024-07-15 18:12:27.482640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.805 [2024-07-15 18:12:27.482647] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.805 [2024-07-15 18:12:27.482823] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.805 [2024-07-15 18:12:27.483000] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.805 [2024-07-15 18:12:27.483007] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.805 [2024-07-15 18:12:27.483013] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.805 [2024-07-15 18:12:27.485839] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.805 [2024-07-15 18:12:27.495356] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.805 [2024-07-15 18:12:27.495782] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.805 [2024-07-15 18:12:27.495798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.805 [2024-07-15 18:12:27.495805] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.805 [2024-07-15 18:12:27.495982] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.805 [2024-07-15 18:12:27.496159] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.805 [2024-07-15 18:12:27.496167] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.805 [2024-07-15 18:12:27.496173] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.805 [2024-07-15 18:12:27.498998] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.805 [2024-07-15 18:12:27.508507] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.805 [2024-07-15 18:12:27.508978] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.805 [2024-07-15 18:12:27.508994] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.805 [2024-07-15 18:12:27.509001] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.805 [2024-07-15 18:12:27.509177] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.805 [2024-07-15 18:12:27.509357] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.805 [2024-07-15 18:12:27.509365] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.805 [2024-07-15 18:12:27.509372] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.805 [2024-07-15 18:12:27.512188] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:33.805 [2024-07-15 18:12:27.521694] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:33.805 [2024-07-15 18:12:27.522146] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:33.805 [2024-07-15 18:12:27.522162] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:33.805 [2024-07-15 18:12:27.522169] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:33.805 [2024-07-15 18:12:27.522350] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:33.805 [2024-07-15 18:12:27.522526] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:33.805 [2024-07-15 18:12:27.522534] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:33.805 [2024-07-15 18:12:27.522540] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:33.805 [2024-07-15 18:12:27.525363] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.063 [2024-07-15 18:12:27.534858] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.063 [2024-07-15 18:12:27.535324] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.063 [2024-07-15 18:12:27.535340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.063 [2024-07-15 18:12:27.535347] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.063 [2024-07-15 18:12:27.535523] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.063 [2024-07-15 18:12:27.535700] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.063 [2024-07-15 18:12:27.535707] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.063 [2024-07-15 18:12:27.535714] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.063 [2024-07-15 18:12:27.538532] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.063 [2024-07-15 18:12:27.548021] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.063 [2024-07-15 18:12:27.548448] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.063 [2024-07-15 18:12:27.548465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.063 [2024-07-15 18:12:27.548471] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.063 [2024-07-15 18:12:27.548648] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.063 [2024-07-15 18:12:27.548824] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.064 [2024-07-15 18:12:27.548832] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.064 [2024-07-15 18:12:27.548838] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.064 [2024-07-15 18:12:27.551658] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.064 [2024-07-15 18:12:27.561149] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.064 [2024-07-15 18:12:27.561569] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.064 [2024-07-15 18:12:27.561586] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.064 [2024-07-15 18:12:27.561595] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.064 [2024-07-15 18:12:27.561771] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.064 [2024-07-15 18:12:27.561948] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.064 [2024-07-15 18:12:27.561956] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.064 [2024-07-15 18:12:27.561962] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.064 [2024-07-15 18:12:27.564783] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.064 [2024-07-15 18:12:27.574290] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.064 [2024-07-15 18:12:27.574728] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.064 [2024-07-15 18:12:27.574744] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.064 [2024-07-15 18:12:27.574751] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.064 [2024-07-15 18:12:27.574927] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.064 [2024-07-15 18:12:27.575105] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.064 [2024-07-15 18:12:27.575113] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.064 [2024-07-15 18:12:27.575119] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.064 [2024-07-15 18:12:27.577946] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.064 [2024-07-15 18:12:27.587444] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.064 [2024-07-15 18:12:27.587884] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.064 [2024-07-15 18:12:27.587900] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.064 [2024-07-15 18:12:27.587908] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.064 [2024-07-15 18:12:27.588084] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.064 [2024-07-15 18:12:27.588264] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.064 [2024-07-15 18:12:27.588273] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.064 [2024-07-15 18:12:27.588279] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.064 [2024-07-15 18:12:27.591097] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.064 [2024-07-15 18:12:27.600594] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.064 [2024-07-15 18:12:27.601082] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.064 [2024-07-15 18:12:27.601097] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.064 [2024-07-15 18:12:27.601104] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.064 [2024-07-15 18:12:27.601284] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.064 [2024-07-15 18:12:27.601460] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.064 [2024-07-15 18:12:27.601472] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.064 [2024-07-15 18:12:27.601478] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.064 [2024-07-15 18:12:27.604304] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.064 [2024-07-15 18:12:27.613630] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.064 [2024-07-15 18:12:27.614013] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.064 [2024-07-15 18:12:27.614029] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.064 [2024-07-15 18:12:27.614035] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.064 [2024-07-15 18:12:27.614212] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.064 [2024-07-15 18:12:27.614394] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.064 [2024-07-15 18:12:27.614403] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.064 [2024-07-15 18:12:27.614409] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.064 [2024-07-15 18:12:27.617231] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.064 [2024-07-15 18:12:27.626727] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.064 [2024-07-15 18:12:27.627118] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.064 [2024-07-15 18:12:27.627134] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.064 [2024-07-15 18:12:27.627140] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.064 [2024-07-15 18:12:27.627321] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.064 [2024-07-15 18:12:27.627502] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.064 [2024-07-15 18:12:27.627509] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.064 [2024-07-15 18:12:27.627516] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.064 [2024-07-15 18:12:27.630337] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.064 [2024-07-15 18:12:27.639831] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.064 [2024-07-15 18:12:27.640303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.064 [2024-07-15 18:12:27.640319] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.064 [2024-07-15 18:12:27.640326] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.064 [2024-07-15 18:12:27.640502] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.064 [2024-07-15 18:12:27.640679] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.064 [2024-07-15 18:12:27.640687] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.064 [2024-07-15 18:12:27.640693] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.064 [2024-07-15 18:12:27.643518] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.064 [2024-07-15 18:12:27.653006] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.064 [2024-07-15 18:12:27.653492] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.064 [2024-07-15 18:12:27.653508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.064 [2024-07-15 18:12:27.653515] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.064 [2024-07-15 18:12:27.653691] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.064 [2024-07-15 18:12:27.653868] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.064 [2024-07-15 18:12:27.653876] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.064 [2024-07-15 18:12:27.653882] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.064 [2024-07-15 18:12:27.656713] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.064 [2024-07-15 18:12:27.666042] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.064 [2024-07-15 18:12:27.666513] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.064 [2024-07-15 18:12:27.666529] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.064 [2024-07-15 18:12:27.666535] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.064 [2024-07-15 18:12:27.666712] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.064 [2024-07-15 18:12:27.666889] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.064 [2024-07-15 18:12:27.666897] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.064 [2024-07-15 18:12:27.666903] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.064 [2024-07-15 18:12:27.669727] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.064 [2024-07-15 18:12:27.679211] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.064 [2024-07-15 18:12:27.679606] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.064 [2024-07-15 18:12:27.679622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.064 [2024-07-15 18:12:27.679629] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.064 [2024-07-15 18:12:27.679805] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.064 [2024-07-15 18:12:27.679982] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.064 [2024-07-15 18:12:27.679990] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.064 [2024-07-15 18:12:27.679996] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.064 [2024-07-15 18:12:27.682820] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.064 [2024-07-15 18:12:27.692312] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.064 [2024-07-15 18:12:27.692780] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.064 [2024-07-15 18:12:27.692795] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.064 [2024-07-15 18:12:27.692802] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.065 [2024-07-15 18:12:27.692982] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.065 [2024-07-15 18:12:27.693159] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.065 [2024-07-15 18:12:27.693167] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.065 [2024-07-15 18:12:27.693173] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.065 [2024-07-15 18:12:27.695996] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.065 [2024-07-15 18:12:27.705488] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.065 [2024-07-15 18:12:27.705957] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.065 [2024-07-15 18:12:27.705973] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.065 [2024-07-15 18:12:27.705980] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.065 [2024-07-15 18:12:27.706157] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.065 [2024-07-15 18:12:27.706337] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.065 [2024-07-15 18:12:27.706346] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.065 [2024-07-15 18:12:27.706352] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.065 [2024-07-15 18:12:27.709171] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.065 [2024-07-15 18:12:27.718671] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.065 [2024-07-15 18:12:27.719140] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.065 [2024-07-15 18:12:27.719156] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.065 [2024-07-15 18:12:27.719163] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.065 [2024-07-15 18:12:27.719345] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.065 [2024-07-15 18:12:27.719523] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.065 [2024-07-15 18:12:27.719531] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.065 [2024-07-15 18:12:27.719537] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.065 [2024-07-15 18:12:27.722358] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.065 [2024-07-15 18:12:27.732050] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.065 [2024-07-15 18:12:27.732535] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.065 [2024-07-15 18:12:27.732552] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.065 [2024-07-15 18:12:27.732559] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.065 [2024-07-15 18:12:27.732736] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.065 [2024-07-15 18:12:27.732914] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.065 [2024-07-15 18:12:27.732923] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.065 [2024-07-15 18:12:27.732937] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.065 [2024-07-15 18:12:27.735761] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.065 [2024-07-15 18:12:27.745101] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.065 [2024-07-15 18:12:27.745501] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.065 [2024-07-15 18:12:27.745517] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.065 [2024-07-15 18:12:27.745524] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.065 [2024-07-15 18:12:27.745700] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.065 [2024-07-15 18:12:27.745877] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.065 [2024-07-15 18:12:27.745885] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.065 [2024-07-15 18:12:27.745891] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.065 [2024-07-15 18:12:27.748716] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.065 [2024-07-15 18:12:27.758206] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.065 [2024-07-15 18:12:27.758658] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.065 [2024-07-15 18:12:27.758674] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.065 [2024-07-15 18:12:27.758681] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.065 [2024-07-15 18:12:27.758857] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.065 [2024-07-15 18:12:27.759033] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.065 [2024-07-15 18:12:27.759041] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.065 [2024-07-15 18:12:27.759047] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.065 [2024-07-15 18:12:27.761868] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.065 [2024-07-15 18:12:27.771367] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.065 [2024-07-15 18:12:27.771818] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.065 [2024-07-15 18:12:27.771835] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.065 [2024-07-15 18:12:27.771842] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.065 [2024-07-15 18:12:27.772018] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.065 [2024-07-15 18:12:27.772195] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.065 [2024-07-15 18:12:27.772203] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.065 [2024-07-15 18:12:27.772209] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.065 [2024-07-15 18:12:27.775034] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.065 18:12:27 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:26:34.065 18:12:27 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@862 -- # return 0 00:26:34.065 18:12:27 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:26:34.065 18:12:27 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@728 -- # xtrace_disable 00:26:34.065 18:12:27 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:26:34.065 [2024-07-15 18:12:27.784530] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.065 [2024-07-15 18:12:27.784960] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.065 [2024-07-15 18:12:27.784976] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.065 [2024-07-15 18:12:27.784983] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.065 [2024-07-15 18:12:27.785160] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.065 [2024-07-15 18:12:27.785343] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.065 [2024-07-15 18:12:27.785351] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.065 [2024-07-15 18:12:27.785357] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.065 [2024-07-15 18:12:27.788183] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.323 [2024-07-15 18:12:27.797699] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.323 [2024-07-15 18:12:27.798161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.323 [2024-07-15 18:12:27.798177] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.323 [2024-07-15 18:12:27.798185] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.323 [2024-07-15 18:12:27.798365] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.323 [2024-07-15 18:12:27.798542] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.323 [2024-07-15 18:12:27.798550] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.323 [2024-07-15 18:12:27.798556] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.323 [2024-07-15 18:12:27.801376] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.323 [2024-07-15 18:12:27.811051] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.323 [2024-07-15 18:12:27.811479] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.323 [2024-07-15 18:12:27.811495] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.323 [2024-07-15 18:12:27.811503] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.323 [2024-07-15 18:12:27.811680] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.323 [2024-07-15 18:12:27.811856] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.323 [2024-07-15 18:12:27.811864] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.323 [2024-07-15 18:12:27.811870] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.323 18:12:27 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:26:34.323 18:12:27 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@17 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:26:34.323 18:12:27 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:34.323 18:12:27 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:26:34.323 [2024-07-15 18:12:27.814699] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.323 [2024-07-15 18:12:27.819302] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:26:34.323 [2024-07-15 18:12:27.824205] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.323 [2024-07-15 18:12:27.824533] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.323 [2024-07-15 18:12:27.824550] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.323 [2024-07-15 18:12:27.824557] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.323 [2024-07-15 18:12:27.824732] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.323 [2024-07-15 18:12:27.824909] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.323 [2024-07-15 18:12:27.824917] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.324 [2024-07-15 18:12:27.824923] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.324 [2024-07-15 18:12:27.827741] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.324 [2024-07-15 18:12:27.837249] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.324 [2024-07-15 18:12:27.837707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.324 [2024-07-15 18:12:27.837722] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.324 [2024-07-15 18:12:27.837729] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.324 [2024-07-15 18:12:27.837906] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.324 [2024-07-15 18:12:27.838084] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.324 [2024-07-15 18:12:27.838091] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.324 [2024-07-15 18:12:27.838098] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.324 [2024-07-15 18:12:27.840921] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.324 18:12:27 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:34.324 18:12:27 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@18 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:26:34.324 18:12:27 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:34.324 18:12:27 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:26:34.324 [2024-07-15 18:12:27.850419] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.324 [2024-07-15 18:12:27.850866] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.324 [2024-07-15 18:12:27.850882] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.324 [2024-07-15 18:12:27.850889] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.324 [2024-07-15 18:12:27.851065] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.324 [2024-07-15 18:12:27.851248] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.324 [2024-07-15 18:12:27.851256] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.324 [2024-07-15 18:12:27.851262] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.324 [2024-07-15 18:12:27.854085] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.324 [2024-07-15 18:12:27.863604] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.324 [2024-07-15 18:12:27.863994] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.324 [2024-07-15 18:12:27.864012] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.324 [2024-07-15 18:12:27.864019] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.324 [2024-07-15 18:12:27.864197] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.324 [2024-07-15 18:12:27.864378] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.324 [2024-07-15 18:12:27.864387] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.324 [2024-07-15 18:12:27.864393] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.324 Malloc0 00:26:34.324 18:12:27 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:34.324 18:12:27 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@19 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:26:34.324 18:12:27 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:34.324 18:12:27 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:26:34.324 [2024-07-15 18:12:27.867208] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.324 [2024-07-15 18:12:27.876695] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.324 [2024-07-15 18:12:27.877064] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:34.324 [2024-07-15 18:12:27.877080] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1553980 with addr=10.0.0.2, port=4420 00:26:34.324 [2024-07-15 18:12:27.877087] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1553980 is same with the state(5) to be set 00:26:34.324 [2024-07-15 18:12:27.877267] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1553980 (9): Bad file descriptor 00:26:34.324 [2024-07-15 18:12:27.877443] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Ctrlr is in error state 00:26:34.324 [2024-07-15 18:12:27.877451] nvme_ctrlr.c:1818:spdk_nvme_ctrlr_reconnect_poll_async: *ERROR*: [nqn.2016-06.io.spdk:cnode1] controller reinitialization failed 00:26:34.324 [2024-07-15 18:12:27.877457] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode1] in failed state. 00:26:34.324 18:12:27 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:34.324 18:12:27 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@20 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:26:34.324 18:12:27 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:34.324 18:12:27 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:26:34.324 [2024-07-15 18:12:27.880279] bdev_nvme.c:2065:_bdev_nvme_reset_ctrlr_complete: *ERROR*: Resetting controller failed. 00:26:34.324 18:12:27 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:34.324 18:12:27 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@21 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:26:34.324 18:12:27 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:34.324 18:12:27 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:26:34.324 [2024-07-15 18:12:27.889199] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:34.324 [2024-07-15 18:12:27.889775] nvme_ctrlr.c:1720:nvme_ctrlr_disconnect: *NOTICE*: [nqn.2016-06.io.spdk:cnode1] resetting controller 00:26:34.324 18:12:27 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:34.324 18:12:27 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@38 -- # wait 739984 00:26:34.324 [2024-07-15 18:12:27.972805] bdev_nvme.c:2067:_bdev_nvme_reset_ctrlr_complete: *NOTICE*: Resetting controller successful. 00:26:44.328 00:26:44.328 Latency(us) 00:26:44.328 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:26:44.328 Job: Nvme1n1 (Core Mask 0x1, workload: verify, depth: 128, IO size: 4096) 00:26:44.328 Verification LBA range: start 0x0 length 0x4000 00:26:44.328 Nvme1n1 : 15.01 8049.09 31.44 12864.36 0.00 6100.70 655.36 16298.52 00:26:44.328 =================================================================================================================== 00:26:44.328 Total : 8049.09 31.44 12864.36 0.00 6100.70 655.36 16298.52 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@39 -- # sync 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@40 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@42 -- # trap - SIGINT SIGTERM EXIT 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- host/bdevperf.sh@44 -- # nvmftestfini 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@488 -- # nvmfcleanup 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@117 -- # sync 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@120 -- # set +e 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@121 -- # for i in {1..20} 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:26:44.328 rmmod nvme_tcp 00:26:44.328 rmmod nvme_fabrics 00:26:44.328 rmmod nvme_keyring 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@124 -- # set -e 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@125 -- # return 0 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@489 -- # '[' -n 740910 ']' 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@490 -- # killprocess 740910 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@948 -- # '[' -z 740910 ']' 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@952 -- # kill -0 740910 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@953 -- # uname 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 740910 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:26:44.328 18:12:36 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:26:44.329 18:12:36 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@966 -- # echo 'killing process with pid 740910' 00:26:44.329 killing process with pid 740910 00:26:44.329 18:12:36 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@967 -- # kill 740910 00:26:44.329 18:12:36 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@972 -- # wait 740910 00:26:44.329 18:12:36 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:26:44.329 18:12:36 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:26:44.329 18:12:36 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:26:44.329 18:12:36 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:26:44.329 18:12:36 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@278 -- # remove_spdk_ns 00:26:44.329 18:12:36 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:44.329 18:12:36 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:26:44.329 18:12:36 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:45.262 18:12:38 nvmf_tcp.nvmf_bdevperf -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:26:45.262 00:26:45.262 real 0m25.451s 00:26:45.262 user 1m2.786s 00:26:45.262 sys 0m5.618s 00:26:45.262 18:12:38 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@1124 -- # xtrace_disable 00:26:45.262 18:12:38 nvmf_tcp.nvmf_bdevperf -- common/autotest_common.sh@10 -- # set +x 00:26:45.262 ************************************ 00:26:45.262 END TEST nvmf_bdevperf 00:26:45.262 ************************************ 00:26:45.262 18:12:38 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:26:45.262 18:12:38 nvmf_tcp -- nvmf/nvmf.sh@123 -- # run_test nvmf_target_disconnect /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/target_disconnect.sh --transport=tcp 00:26:45.262 18:12:38 nvmf_tcp -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:26:45.262 18:12:38 nvmf_tcp -- common/autotest_common.sh@1105 -- # xtrace_disable 00:26:45.262 18:12:38 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:26:45.262 ************************************ 00:26:45.262 START TEST nvmf_target_disconnect 00:26:45.262 ************************************ 00:26:45.262 18:12:38 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/target_disconnect.sh --transport=tcp 00:26:45.522 * Looking for test storage... 00:26:45.522 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- host/target_disconnect.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@7 -- # uname -s 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- paths/export.sh@5 -- # export PATH 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@47 -- # : 0 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@51 -- # have_pci_nics=0 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- host/target_disconnect.sh@11 -- # PLUGIN_DIR=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/app/fio/nvme 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- host/target_disconnect.sh@13 -- # MALLOC_BDEV_SIZE=64 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- host/target_disconnect.sh@14 -- # MALLOC_BLOCK_SIZE=512 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- host/target_disconnect.sh@69 -- # nvmftestinit 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@448 -- # prepare_net_devs 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@410 -- # local -g is_hw=no 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@412 -- # remove_spdk_ns 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@285 -- # xtrace_disable 00:26:45.522 18:12:39 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@10 -- # set +x 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@291 -- # pci_devs=() 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@291 -- # local -a pci_devs 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@292 -- # pci_net_devs=() 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@293 -- # pci_drivers=() 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@293 -- # local -A pci_drivers 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@295 -- # net_devs=() 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@295 -- # local -ga net_devs 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@296 -- # e810=() 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@296 -- # local -ga e810 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@297 -- # x722=() 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@297 -- # local -ga x722 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@298 -- # mlx=() 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@298 -- # local -ga mlx 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:26:50.816 Found 0000:86:00.0 (0x8086 - 0x159b) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:26:50.816 Found 0000:86:00.1 (0x8086 - 0x159b) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@390 -- # [[ up == up ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:26:50.816 Found net devices under 0000:86:00.0: cvl_0_0 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@390 -- # [[ up == up ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:26:50.816 Found net devices under 0000:86:00.1: cvl_0_1 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@414 -- # is_hw=yes 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:26:50.816 18:12:43 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:26:50.816 18:12:44 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:26:50.816 18:12:44 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:26:50.816 18:12:44 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:26:50.816 18:12:44 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:26:50.816 18:12:44 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:26:50.817 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:26:50.817 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.271 ms 00:26:50.817 00:26:50.817 --- 10.0.0.2 ping statistics --- 00:26:50.817 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:50.817 rtt min/avg/max/mdev = 0.271/0.271/0.271/0.000 ms 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:26:50.817 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:26:50.817 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.262 ms 00:26:50.817 00:26:50.817 --- 10.0.0.1 ping statistics --- 00:26:50.817 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:26:50.817 rtt min/avg/max/mdev = 0.262/0.262/0.262/0.000 ms 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@422 -- # return 0 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- host/target_disconnect.sh@70 -- # run_test nvmf_target_disconnect_tc1 nvmf_target_disconnect_tc1 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@1105 -- # xtrace_disable 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@10 -- # set +x 00:26:50.817 ************************************ 00:26:50.817 START TEST nvmf_target_disconnect_tc1 00:26:50.817 ************************************ 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@1123 -- # nvmf_target_disconnect_tc1 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- host/target_disconnect.sh@32 -- # NOT /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@648 -- # local es=0 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@650 -- # valid_exec_arg /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@636 -- # local arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@640 -- # type -t /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@642 -- # type -P /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@642 -- # arg=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@642 -- # [[ -x /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect ]] 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@651 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:26:50.817 EAL: No free 2048 kB hugepages reported on node 1 00:26:50.817 [2024-07-15 18:12:44.322210] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:50.817 [2024-07-15 18:12:44.322332] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1ecae60 with addr=10.0.0.2, port=4420 00:26:50.817 [2024-07-15 18:12:44.322386] nvme_tcp.c:2711:nvme_tcp_ctrlr_construct: *ERROR*: failed to create admin qpair 00:26:50.817 [2024-07-15 18:12:44.322420] nvme.c: 830:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:26:50.817 [2024-07-15 18:12:44.322440] nvme.c: 913:spdk_nvme_probe: *ERROR*: Create probe context failed 00:26:50.817 spdk_nvme_probe() failed for transport address '10.0.0.2' 00:26:50.817 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect: errors occurred 00:26:50.817 Initializing NVMe Controllers 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@651 -- # es=1 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:26:50.817 00:26:50.817 real 0m0.097s 00:26:50.817 user 0m0.045s 00:26:50.817 sys 0m0.052s 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@1124 -- # xtrace_disable 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc1 -- common/autotest_common.sh@10 -- # set +x 00:26:50.817 ************************************ 00:26:50.817 END TEST nvmf_target_disconnect_tc1 00:26:50.817 ************************************ 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@1142 -- # return 0 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- host/target_disconnect.sh@71 -- # run_test nvmf_target_disconnect_tc2 nvmf_target_disconnect_tc2 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@1105 -- # xtrace_disable 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@10 -- # set +x 00:26:50.817 ************************************ 00:26:50.817 START TEST nvmf_target_disconnect_tc2 00:26:50.817 ************************************ 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@1123 -- # nvmf_target_disconnect_tc2 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@37 -- # disconnect_init 10.0.0.2 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@17 -- # nvmfappstart -m 0xF0 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@722 -- # xtrace_disable 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@481 -- # nvmfpid=745985 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@482 -- # waitforlisten 745985 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF0 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@829 -- # '[' -z 745985 ']' 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@834 -- # local max_retries=100 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:50.817 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@838 -- # xtrace_disable 00:26:50.817 18:12:44 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:50.817 [2024-07-15 18:12:44.452013] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:26:50.817 [2024-07-15 18:12:44.452060] [ DPDK EAL parameters: nvmf -c 0xF0 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:50.817 EAL: No free 2048 kB hugepages reported on node 1 00:26:50.817 [2024-07-15 18:12:44.519634] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:26:51.077 [2024-07-15 18:12:44.597814] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:26:51.077 [2024-07-15 18:12:44.597851] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:26:51.077 [2024-07-15 18:12:44.597858] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:26:51.077 [2024-07-15 18:12:44.597864] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:26:51.077 [2024-07-15 18:12:44.597868] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:26:51.077 [2024-07-15 18:12:44.597926] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 5 00:26:51.077 [2024-07-15 18:12:44.598034] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 6 00:26:51.077 [2024-07-15 18:12:44.598150] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 4 00:26:51.077 [2024-07-15 18:12:44.598152] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 7 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@862 -- # return 0 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@728 -- # xtrace_disable 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:51.645 Malloc0 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:51.645 [2024-07-15 18:12:45.315389] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:51.645 [2024-07-15 18:12:45.340385] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@42 -- # reconnectpid=746092 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@44 -- # sleep 2 00:26:51.645 18:12:45 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@40 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/reconnect -q 32 -o 4096 -w randrw -M 50 -t 10 -c 0xF -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:26:51.904 EAL: No free 2048 kB hugepages reported on node 1 00:26:53.823 18:12:47 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@45 -- # kill -9 745985 00:26:53.823 18:12:47 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@47 -- # sleep 2 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Write completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Write completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Write completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Write completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Write completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Write completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Write completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 [2024-07-15 18:12:47.367455] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 4 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Write completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Write completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Write completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Write completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Write completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Write completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Write completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.823 Read completed with error (sct=0, sc=8) 00:26:53.823 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 [2024-07-15 18:12:47.367657] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 [2024-07-15 18:12:47.367849] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Write completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 Read completed with error (sct=0, sc=8) 00:26:53.824 starting I/O failed 00:26:53.824 [2024-07-15 18:12:47.368043] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:26:53.824 [2024-07-15 18:12:47.368276] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.824 [2024-07-15 18:12:47.368292] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.824 qpair failed and we were unable to recover it. 00:26:53.824 [2024-07-15 18:12:47.368433] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.824 [2024-07-15 18:12:47.368442] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.824 qpair failed and we were unable to recover it. 00:26:53.824 [2024-07-15 18:12:47.368580] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.824 [2024-07-15 18:12:47.368590] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.824 qpair failed and we were unable to recover it. 00:26:53.824 [2024-07-15 18:12:47.368797] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.824 [2024-07-15 18:12:47.368807] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.824 qpair failed and we were unable to recover it. 00:26:53.824 [2024-07-15 18:12:47.368987] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.824 [2024-07-15 18:12:47.368997] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.824 qpair failed and we were unable to recover it. 00:26:53.824 [2024-07-15 18:12:47.369162] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.824 [2024-07-15 18:12:47.369173] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.824 qpair failed and we were unable to recover it. 00:26:53.824 [2024-07-15 18:12:47.369335] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.824 [2024-07-15 18:12:47.369345] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.824 qpair failed and we were unable to recover it. 00:26:53.824 [2024-07-15 18:12:47.369484] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.824 [2024-07-15 18:12:47.369494] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.824 qpair failed and we were unable to recover it. 00:26:53.824 [2024-07-15 18:12:47.369669] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.824 [2024-07-15 18:12:47.369680] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.824 qpair failed and we were unable to recover it. 00:26:53.824 [2024-07-15 18:12:47.369839] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.824 [2024-07-15 18:12:47.369849] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.824 qpair failed and we were unable to recover it. 00:26:53.824 [2024-07-15 18:12:47.370054] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.824 [2024-07-15 18:12:47.370064] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.824 qpair failed and we were unable to recover it. 00:26:53.824 [2024-07-15 18:12:47.370232] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.824 [2024-07-15 18:12:47.370242] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.824 qpair failed and we were unable to recover it. 00:26:53.824 [2024-07-15 18:12:47.370443] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.824 [2024-07-15 18:12:47.370472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.824 qpair failed and we were unable to recover it. 00:26:53.824 [2024-07-15 18:12:47.370704] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.370734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.371131] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.371161] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.371390] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.371421] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.371597] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.371627] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.371789] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.371818] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.372120] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.372158] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.372407] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.372421] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.372623] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.372637] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.372820] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.372834] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.373083] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.373112] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.373266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.373297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.373478] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.373508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.373668] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.373697] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.373976] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.374021] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.374270] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.374281] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.374484] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.374494] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.374679] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.374689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.374856] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.374866] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.375047] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.375057] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.375167] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.375177] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.375417] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.375427] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.375608] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.375619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.375738] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.375747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.375974] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.375983] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.376176] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.376185] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.376354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.376363] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.376491] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.376505] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.376619] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.376629] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.376888] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.376898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.377081] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.377091] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.377267] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.377278] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.377484] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.377495] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.377633] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.377642] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.825 [2024-07-15 18:12:47.377741] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.825 [2024-07-15 18:12:47.377751] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.825 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.377945] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.377954] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.378157] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.378167] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.378400] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.378411] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.378639] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.378648] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.378772] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.378782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.378989] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.378999] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.379181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.379191] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.379369] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.379379] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.379607] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.379617] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.379743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.379752] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.379942] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.379971] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.380174] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.380204] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.380387] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.380417] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.380590] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.380619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.380837] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.380866] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.381067] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.381096] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.381301] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.381311] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.381548] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.381577] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.381737] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.381766] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.382088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.382118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.382405] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.382434] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.382591] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.382621] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.382786] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.382815] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.383103] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.383132] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.383303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.383313] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.383433] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.383443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.383623] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.383633] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.383793] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.383803] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.384044] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.384054] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.384269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.384279] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.384460] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.384471] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.384651] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.384661] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.384880] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.384892] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.385163] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.385173] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.385368] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.385378] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.385503] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.385513] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.385696] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.385706] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.385824] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.385834] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.386086] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.386096] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.386310] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.386321] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.826 qpair failed and we were unable to recover it. 00:26:53.826 [2024-07-15 18:12:47.386504] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.826 [2024-07-15 18:12:47.386514] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.386625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.386635] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.386807] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.386817] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.387025] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.387034] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.387266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.387276] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.387395] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.387405] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.387535] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.387545] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.387735] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.387745] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.387851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.387861] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.387972] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.387981] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.388147] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.388157] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.388341] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.388351] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.388525] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.388535] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.388655] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.388665] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.388840] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.388850] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.389045] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.389055] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.389244] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.389254] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.389396] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.389406] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.389600] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.389629] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.389946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.389976] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.390280] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.390291] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.390483] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.390494] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.390688] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.390718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.390960] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.390990] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.391280] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.391311] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.391482] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.391511] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.391686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.391715] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.392002] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.392012] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.392229] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.392239] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.392438] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.392448] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.392609] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.392619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.392746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.392756] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.393056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.393068] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.393234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.393244] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.393478] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.393508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.393733] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.393764] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.394031] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.394060] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.394310] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.394340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.394631] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.394661] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.394905] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.394934] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.827 [2024-07-15 18:12:47.395223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.827 [2024-07-15 18:12:47.395294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.827 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.395616] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.395645] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.395801] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.395830] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.396080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.396109] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.396313] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.396323] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.396583] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.396613] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.396880] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.396909] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.397134] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.397163] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.397380] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.397410] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.397706] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.397736] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.398035] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.398065] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.398297] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.398307] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.398417] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.398426] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.398561] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.398571] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.398665] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.398674] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.398930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.398940] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.399050] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.399060] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.399310] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.399320] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.399440] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.399450] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.399585] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.399595] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.399753] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.399763] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.399875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.399885] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.400053] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.400062] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.400242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.400252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.400512] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.400522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.400661] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.400670] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.400964] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.400973] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.401175] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.401186] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.401346] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.401357] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.401602] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.401632] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.401944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.401973] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.402127] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.402164] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.828 [2024-07-15 18:12:47.402339] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.828 [2024-07-15 18:12:47.402351] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.828 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.402514] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.402524] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.402705] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.402714] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.402980] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.402990] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.403172] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.403181] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.403412] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.403443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.403681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.403710] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.404001] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.404030] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.404274] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.404304] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.404478] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.404507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.404730] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.404759] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.404977] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.405007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.405161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.405171] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.405363] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.405393] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.405694] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.405724] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.406082] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.406111] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.406370] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.406400] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.406732] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.406762] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.406916] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.406944] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.407114] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.407143] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.407438] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.407468] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.407629] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.407658] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.407928] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.407957] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.408265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.408295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.408590] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.408619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.408891] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.408921] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.409126] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.409155] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.409460] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.409528] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.409824] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.409857] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.410160] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.410174] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.410363] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.410378] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.410556] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.410570] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.410805] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.410835] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.411075] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.411105] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.411331] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.411361] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.411610] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.411640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.411902] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.411931] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.412271] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.412302] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.412525] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.412555] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.412757] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.412787] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.413003] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.413042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.829 [2024-07-15 18:12:47.413187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.829 [2024-07-15 18:12:47.413216] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.829 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.413542] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.413572] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.413791] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.413821] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.414144] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.414173] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.414355] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.414386] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.414662] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.414692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.414858] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.414888] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.415206] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.415248] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.415422] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.415452] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.415612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.415642] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.415867] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.415896] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.416181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.416211] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.416449] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.416463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.416734] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.416748] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.416880] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.416894] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.417035] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.417048] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.417159] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.417172] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.417309] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.417323] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.417539] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.417553] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.417685] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.417699] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.417993] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.418007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.418218] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.418255] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.418525] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.418555] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.418781] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.418812] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.418961] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.418975] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.419213] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.419257] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.419431] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.419461] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.419631] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.419661] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.419829] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.419858] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.420081] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.420111] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.420406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.420437] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.420608] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.420637] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.420927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.420956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.421195] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.421216] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.421389] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.421403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.421599] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.421629] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.421786] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.421815] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.422031] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.422061] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.422282] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.422313] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.422526] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.422555] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.422790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.422820] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.830 [2024-07-15 18:12:47.423088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.830 [2024-07-15 18:12:47.423118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.830 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.423346] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.423376] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.423534] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.423563] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.423721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.423751] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.424038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.424067] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.424381] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.424412] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.424578] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.424607] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.424899] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.424929] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.425194] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.425207] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.425352] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.425366] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.425648] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.425662] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.425768] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.425782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.425992] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.426005] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.426165] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.426179] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.426389] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.426403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.426514] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.426528] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.426795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.426808] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.427002] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.427017] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.427235] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.427249] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.427374] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.427388] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.427558] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.427572] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.427751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.427764] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.428031] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.428046] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.428223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.428245] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.428511] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.428525] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.428769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.428788] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.428999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.429013] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.429217] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.429237] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.429428] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.429442] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.429577] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.429592] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.429727] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.429741] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.429847] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.429861] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.430031] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.430044] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.430257] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.430271] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.430462] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.430476] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.430611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.430624] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.430919] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.430948] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.431219] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.431269] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.431447] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.431476] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.431752] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.431782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.432093] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.432122] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.831 [2024-07-15 18:12:47.432413] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.831 [2024-07-15 18:12:47.432444] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.831 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.432659] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.432688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.432830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.432859] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.433067] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.433096] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.433372] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.433386] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.433624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.433638] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.433810] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.433824] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.434094] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.434107] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.434359] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.434390] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.434551] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.434581] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.434796] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.434825] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.435049] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.435062] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.435250] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.435263] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.435500] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.435514] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.435684] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.435697] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.435901] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.435930] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.436199] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.436235] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.436558] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.436587] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.436885] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.436915] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.437188] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.437217] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.437544] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.437574] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.437903] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.437939] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.438119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.438133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.438320] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.438350] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.438573] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.438607] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.438755] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.438784] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.439111] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.439141] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.439401] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.439432] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.439658] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.439688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.439927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.439955] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.440244] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.440274] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.440544] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.440558] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.440752] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.440767] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.441029] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.441043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.832 qpair failed and we were unable to recover it. 00:26:53.832 [2024-07-15 18:12:47.441319] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.832 [2024-07-15 18:12:47.441349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.441602] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.441632] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.441930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.441959] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.442169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.442199] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.442425] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.442439] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.442695] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.442709] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.442926] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.442940] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.443203] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.443217] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.443422] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.443435] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.443723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.443737] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.444038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.444052] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.444265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.444281] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.444459] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.444473] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.444740] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.444769] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.445085] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.445115] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.445347] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.445361] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.445548] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.445562] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.445830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.445859] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.446080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.446094] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.446354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.446368] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.446537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.446551] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.446754] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.446783] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.447004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.447033] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.447269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.447301] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.447586] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.447600] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.447787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.447801] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.448087] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.448100] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.448276] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.448291] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.448535] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.448564] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.448812] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.448842] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.449183] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.449219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.449493] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.449522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.449841] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.449870] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.450176] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.450206] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.450494] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.450508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.450689] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.450703] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.450932] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.450947] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.451065] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.451079] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.451347] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.451363] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.451621] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.451634] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.451897] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.833 [2024-07-15 18:12:47.451911] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.833 qpair failed and we were unable to recover it. 00:26:53.833 [2024-07-15 18:12:47.452171] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.452186] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.452373] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.452387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.452692] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.452706] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.452842] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.452855] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.453095] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.453109] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.453388] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.453419] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.453686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.453716] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.454030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.454059] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.454351] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.454381] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.454599] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.454628] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.454799] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.454829] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.455033] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.455062] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.455347] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.455378] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.455605] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.455634] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.455914] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.455943] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.456254] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.456284] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.456565] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.456594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.456764] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.456793] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.457023] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.457053] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.457260] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.457291] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.457520] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.457548] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.457776] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.457806] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.458018] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.458032] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.458243] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.458257] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.458473] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.458487] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.458661] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.458674] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.458864] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.458893] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.459164] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.459192] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.459524] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.459554] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.459780] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.459815] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.460102] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.460115] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.460236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.460250] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.460370] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.460384] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.834 [2024-07-15 18:12:47.460570] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.834 [2024-07-15 18:12:47.460584] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.834 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.460776] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.460790] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.460984] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.460997] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.461205] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.461243] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.461414] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.461444] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.461740] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.461769] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.461982] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.462011] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.462295] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.462325] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.462547] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.462575] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.462875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.462904] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.463201] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.463256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.463532] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.463546] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.463750] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.463763] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.463977] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.463991] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.464235] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.464249] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.464498] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.464528] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.464824] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.464854] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.465124] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.465162] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.465481] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.465495] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.465743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.465757] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.465948] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.465962] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.466082] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.466096] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.466283] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.466297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.466565] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.466579] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.466838] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.466858] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.467030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.467045] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.467213] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.467235] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.467492] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.467521] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.467764] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.467793] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.467999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.468028] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.468325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.468340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.468465] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.468479] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.468743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.468757] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.468993] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.469007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.469264] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.469294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.469567] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.469597] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.469830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.469868] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.470165] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.470179] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.470416] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.470430] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.470670] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.470684] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.470874] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.835 [2024-07-15 18:12:47.470904] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.835 qpair failed and we were unable to recover it. 00:26:53.835 [2024-07-15 18:12:47.471110] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.471139] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.471344] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.471375] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.471550] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.471564] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.471751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.471781] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.472106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.472136] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.472418] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.472463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.472738] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.472768] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.473060] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.473090] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.473317] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.473347] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.473625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.473640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.473902] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.473915] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.474096] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.474110] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.474378] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.474392] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.474529] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.474543] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.474786] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.474800] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.475014] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.475027] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.475270] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.475302] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.475518] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.475547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.475840] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.475869] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.476165] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.476179] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.476461] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.476475] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.476739] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.476753] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.476884] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.476898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.477067] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.477106] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.477379] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.477409] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.477683] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.477712] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.478012] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.478041] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.478385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.478416] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.478665] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.478694] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.478991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.479020] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.479326] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.479357] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.479676] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.479705] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.479988] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.480017] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.480262] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.480301] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.836 [2024-07-15 18:12:47.480583] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.836 [2024-07-15 18:12:47.480597] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.836 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.480796] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.480813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.481073] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.481103] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.481385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.481420] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.481659] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.481672] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.481935] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.481949] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.482203] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.482216] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.482463] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.482477] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.482669] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.482682] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.482877] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.482892] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.483157] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.483172] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.483362] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.483376] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.483568] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.483597] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.483756] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.483786] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.483995] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.484024] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.484348] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.484363] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.484578] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.484592] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.484846] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.484860] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.485058] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.485071] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.485275] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.485290] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.485476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.485489] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.485703] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.485733] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.486019] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.486049] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.486350] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.486381] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.486681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.486710] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.487014] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.487043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.487343] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.487373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.487614] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.487643] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.487993] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.488022] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.488277] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.488291] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.488499] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.488527] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.488752] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.488782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.489084] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.489114] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.489396] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.489426] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.489720] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.489749] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.490051] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.490080] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.490379] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.490393] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.490589] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.490603] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.490867] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.490881] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.491173] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.491187] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.491433] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.491448] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.837 qpair failed and we were unable to recover it. 00:26:53.837 [2024-07-15 18:12:47.491686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.837 [2024-07-15 18:12:47.491702] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.491966] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.491979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.492216] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.492235] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.492533] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.492546] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.492816] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.492829] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.493088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.493102] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.493394] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.493408] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.493650] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.493664] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.493848] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.493861] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.494125] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.494138] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.494323] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.494337] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.494512] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.494526] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.494661] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.494674] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.494858] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.494871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.495175] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.495205] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.495453] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.495484] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.495693] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.495723] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.496042] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.496072] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.496359] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.496390] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.496650] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.496680] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.496991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.497028] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.497300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.497314] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.497441] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.497455] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.497676] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.497705] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.497857] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.497887] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.498098] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.498127] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.498403] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.498434] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.498807] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.498879] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.499202] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.499251] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.499567] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.499581] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.499821] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.499836] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.500047] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.500060] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.500359] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.500374] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.500623] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.500637] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.500824] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.500837] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.501102] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.501115] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.501412] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.501426] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.501672] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.501686] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.501924] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.501938] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.502064] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.502078] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.502270] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.502304] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.838 [2024-07-15 18:12:47.502594] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.838 [2024-07-15 18:12:47.502624] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.838 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.502784] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.502814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.503021] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.503050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.503343] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.503383] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.503654] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.503668] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.503795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.503809] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.504049] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.504063] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.504315] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.504329] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.504582] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.504596] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.504855] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.504869] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.505137] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.505151] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.505255] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.505268] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.505458] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.505472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.505779] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.505793] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.506074] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.506088] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.506279] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.506294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.506491] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.506505] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.506769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.506782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.506966] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.506979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.507257] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.507271] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.507560] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.507573] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.507704] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.507718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.507920] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.507934] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.508103] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.508117] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.508394] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.508425] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.508745] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.508775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.509045] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.509075] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.509315] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.509347] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.509636] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.509649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.509922] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.509951] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.510290] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.510321] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.510605] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.510619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.510886] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.510900] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.511032] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.511046] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.511270] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.511301] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.511472] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.511502] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.511725] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.511755] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.512054] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.512084] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.512238] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.512252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.512515] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.512545] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.512839] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.839 [2024-07-15 18:12:47.512874] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.839 qpair failed and we were unable to recover it. 00:26:53.839 [2024-07-15 18:12:47.513179] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.513210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.513425] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.513455] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.513750] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.513780] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.513985] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.514015] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.514237] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.514252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.514515] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.514529] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.514699] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.514713] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.514975] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.515005] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.515236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.515268] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.515557] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.515571] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.515819] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.515833] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.516077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.516091] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.516219] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.516238] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.516530] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.516544] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.516735] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.516748] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.516947] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.516961] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.517202] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.517216] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.517482] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.517497] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.517747] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.517760] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.518004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.518018] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.518222] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.518249] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.518525] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.518539] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.518800] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.518814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.519093] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.519107] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.519406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.519438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.519758] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.519788] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.520067] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.520102] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.520419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.520449] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.520731] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.520761] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.520994] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.521024] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.521230] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.521244] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.521494] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.521508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.521646] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.521659] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.521851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.521865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.522128] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.522142] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.522377] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.522406] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.522649] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.522679] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.522883] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.522913] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.523241] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.523272] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.523564] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.523594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.523896] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.523926] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.840 [2024-07-15 18:12:47.524253] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.840 [2024-07-15 18:12:47.524284] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.840 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.524603] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.524633] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.524883] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.524913] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.525251] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.525282] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.525563] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.525593] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.525885] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.525915] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.526155] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.526185] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.526400] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.526415] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.526667] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.526681] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.526874] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.526887] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.527078] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.527092] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.527358] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.527373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.527593] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.527607] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.527787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.527800] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.527937] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.527970] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.528115] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.528144] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.528420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.528450] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.528767] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.528781] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.529045] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.529059] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.529241] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.529255] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.529446] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.529460] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.529657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.529687] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.529962] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.529992] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.530316] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.530347] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.530570] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.530584] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.530851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.530865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.531179] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.531193] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.531332] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.531346] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.531535] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.531548] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.531862] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.531891] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.532125] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.532155] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.532398] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.532412] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.532667] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.532697] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.532978] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.533007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.533251] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.533282] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.533580] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.533609] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.841 [2024-07-15 18:12:47.533871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.841 [2024-07-15 18:12:47.533901] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.841 qpair failed and we were unable to recover it. 00:26:53.842 [2024-07-15 18:12:47.534182] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.842 [2024-07-15 18:12:47.534213] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.842 qpair failed and we were unable to recover it. 00:26:53.842 [2024-07-15 18:12:47.534449] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.842 [2024-07-15 18:12:47.534479] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.842 qpair failed and we were unable to recover it. 00:26:53.842 [2024-07-15 18:12:47.534616] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.842 [2024-07-15 18:12:47.534630] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.842 qpair failed and we were unable to recover it. 00:26:53.842 [2024-07-15 18:12:47.534878] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.842 [2024-07-15 18:12:47.534908] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.842 qpair failed and we were unable to recover it. 00:26:53.842 [2024-07-15 18:12:47.535212] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.842 [2024-07-15 18:12:47.535251] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.842 qpair failed and we were unable to recover it. 00:26:53.842 [2024-07-15 18:12:47.535540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.842 [2024-07-15 18:12:47.535571] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.842 qpair failed and we were unable to recover it. 00:26:53.842 [2024-07-15 18:12:47.535830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.842 [2024-07-15 18:12:47.535844] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.842 qpair failed and we were unable to recover it. 00:26:53.842 [2024-07-15 18:12:47.536036] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.842 [2024-07-15 18:12:47.536050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.842 qpair failed and we were unable to recover it. 00:26:53.842 [2024-07-15 18:12:47.536336] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.842 [2024-07-15 18:12:47.536350] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.842 qpair failed and we were unable to recover it. 00:26:53.842 [2024-07-15 18:12:47.536539] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.842 [2024-07-15 18:12:47.536553] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.842 qpair failed and we were unable to recover it. 00:26:53.842 [2024-07-15 18:12:47.536750] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.842 [2024-07-15 18:12:47.536763] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.842 qpair failed and we were unable to recover it. 00:26:53.842 [2024-07-15 18:12:47.537004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:53.842 [2024-07-15 18:12:47.537034] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:53.842 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.537388] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.537421] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.537695] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.537709] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.538000] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.538014] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.538187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.538200] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.538398] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.538416] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.538613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.538643] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.538922] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.538952] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.539169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.539199] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.539509] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.539540] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.539768] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.539798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.540111] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.540149] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.540447] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.540461] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.540669] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.540683] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.540884] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.540898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.541092] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.541105] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.541325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.541340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.541536] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.541550] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.541814] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.541827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.542030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.542060] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.542266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.542297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.542524] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.542554] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.542716] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.118 [2024-07-15 18:12:47.542745] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.118 qpair failed and we were unable to recover it. 00:26:54.118 [2024-07-15 18:12:47.543038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.543068] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.543217] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.543255] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.543466] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.543496] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.543698] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.543712] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.543956] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.543970] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.544247] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.544262] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.544529] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.544543] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.544799] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.544813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.545056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.545070] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.545331] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.545348] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.545545] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.545560] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.545827] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.545841] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.546094] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.546108] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.546346] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.546360] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.546550] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.546564] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.546829] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.546843] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.547108] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.547122] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.547390] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.547419] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.547669] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.547700] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.547921] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.547952] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.548254] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.548285] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.548529] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.548559] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.548766] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.548780] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.548971] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.548985] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.549232] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.549248] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.549511] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.549526] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.549787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.549801] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.550058] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.550072] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.550315] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.550331] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.550612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.550625] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.550895] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.550909] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.551080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.551094] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.551215] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.551234] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.551524] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.551553] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.551881] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.551911] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.552206] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.552259] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.552547] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.552582] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.552872] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.552902] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.553148] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.553177] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.553533] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.553565] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.553863] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.553892] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.554198] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.554240] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.554562] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.554595] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.554875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.554889] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.555126] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.555140] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.555336] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.555350] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.555584] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.555614] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.555822] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.555851] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.556088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.556118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.556261] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.556292] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.556592] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.556622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.556925] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.556955] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.557264] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.557295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.557518] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.557548] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.557848] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.557878] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.558104] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.558134] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.558413] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.558443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.119 qpair failed and we were unable to recover it. 00:26:54.119 [2024-07-15 18:12:47.558750] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.119 [2024-07-15 18:12:47.558764] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.558955] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.558969] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.559209] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.559229] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.559424] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.559438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.559645] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.559659] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.559914] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.559944] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.560247] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.560279] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.560510] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.560524] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.560790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.560804] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.561072] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.561087] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.561332] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.561347] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.561638] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.561651] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.561917] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.561931] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.562200] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.562214] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.562464] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.562478] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.562747] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.562761] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.562947] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.562961] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.563238] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.563252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.563434] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.563448] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.563723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.563753] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.563933] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.563964] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.564265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.564302] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.564545] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.564559] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.564825] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.564839] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.565109] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.565123] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.565367] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.565382] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.565644] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.565658] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.565913] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.565952] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.566252] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.566282] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.566585] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.566615] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.566845] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.566875] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.567105] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.567135] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.567345] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.567360] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.567586] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.567616] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.567872] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.567902] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.568210] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.568264] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.568564] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.568595] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.568897] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.568927] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.569240] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.569272] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.569504] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.569534] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.569883] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.569913] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.570196] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.570237] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.570537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.570567] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.570790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.570820] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.571113] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.571144] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.571319] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.571350] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.571580] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.571617] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.571873] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.571889] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.572078] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.572092] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.572279] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.572293] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.572506] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.572536] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.572860] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.572889] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.573165] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.573194] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.573513] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.573544] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.573833] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.573863] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.574175] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.574205] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.574413] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.120 [2024-07-15 18:12:47.574428] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.120 qpair failed and we were unable to recover it. 00:26:54.120 [2024-07-15 18:12:47.574621] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.574635] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.574856] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.574869] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.575008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.575022] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.575278] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.575308] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.575549] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.575579] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.575802] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.575832] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.576129] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.576159] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.576387] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.576418] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.576629] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.576659] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.576961] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.576975] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.577239] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.577253] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.577545] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.577559] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.577737] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.577751] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.578042] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.578056] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.578306] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.578321] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.578535] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.578549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.578801] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.578815] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.579064] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.579081] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.579342] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.579357] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.579546] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.579560] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.579875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.579905] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.580161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.580190] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.580533] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.580564] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.580735] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.580764] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.581023] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.581052] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.581408] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.581438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.581671] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.581686] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.581864] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.581878] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.582152] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.582181] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.582480] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.582511] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.582758] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.582772] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.583055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.583069] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.583268] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.583284] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.583501] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.583515] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.583765] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.583795] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.584030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.584059] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.584280] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.584312] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.584593] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.584623] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.584934] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.584964] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.585143] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.585172] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.585442] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.585474] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.585801] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.585815] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.586107] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.586121] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.586405] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.586419] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.586613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.586627] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.586907] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.586921] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.587199] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.587213] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.121 [2024-07-15 18:12:47.587493] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.121 [2024-07-15 18:12:47.587507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.121 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.587771] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.587784] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.588057] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.588071] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.588295] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.588309] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.588577] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.588592] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.588775] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.588790] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.589039] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.589053] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.589238] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.589253] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.589481] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.589512] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.589838] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.589868] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.590144] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.590174] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.590486] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.590502] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.590799] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.590814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.591039] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.591054] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.591347] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.591362] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.591490] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.591505] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.591781] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.591811] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.592103] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.592133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.592369] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.592400] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.592641] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.592672] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.592954] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.592968] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.593242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.593273] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.593510] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.593540] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.593765] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.593804] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.594047] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.594077] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.594367] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.594398] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.594685] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.594715] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.595028] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.595059] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.595350] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.595382] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.595607] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.595637] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.595948] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.595979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.596145] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.596175] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.596523] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.596555] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.596707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.596721] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.596917] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.596932] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.597120] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.597150] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.597384] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.597415] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.597702] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.597731] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.597970] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.598005] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.598220] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.598259] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.598507] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.598521] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.598721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.598736] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.599005] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.599019] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.599204] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.599218] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.599401] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.599415] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.599688] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.599719] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.599928] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.599959] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.600172] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.600202] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.600545] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.600576] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.600888] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.600902] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.601198] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.601212] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.601476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.601491] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.601755] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.601769] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.602020] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.602034] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.602251] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.602266] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.602516] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.602530] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.602829] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.602844] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.603056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.122 [2024-07-15 18:12:47.603070] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.122 qpair failed and we were unable to recover it. 00:26:54.122 [2024-07-15 18:12:47.603311] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.603326] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.603588] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.603602] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.603859] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.603873] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.604124] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.604138] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.604399] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.604414] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.604595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.604610] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.604885] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.604914] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.605244] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.605294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.605488] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.605502] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.605682] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.605696] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.605892] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.605914] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.606176] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.606206] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.606550] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.606580] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.606796] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.606811] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.607060] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.607074] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.607278] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.607293] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.607514] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.607528] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.607706] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.607739] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.608039] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.608069] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.608297] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.608330] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.608641] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.608670] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.608964] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.608993] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.609304] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.609336] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.609638] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.609667] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.609967] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.609981] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.610298] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.610312] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.610572] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.610586] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.610848] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.610862] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.611062] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.611076] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.611281] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.611295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.611428] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.611442] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.611662] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.611676] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.611948] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.611962] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.612246] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.612261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.612462] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.612479] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.612755] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.612769] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.613056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.613071] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.613251] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.613266] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.613564] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.613579] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.613778] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.613792] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.614062] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.614076] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.614374] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.614389] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.614644] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.614658] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.614935] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.614949] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.615230] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.615245] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.615509] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.615523] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.615772] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.615786] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.616040] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.616054] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.616320] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.616335] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.616515] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.616529] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.616746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.616775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.617089] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.617119] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.617452] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.617482] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.617694] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.617723] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.618039] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.618053] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.618341] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.618356] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.618579] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.123 [2024-07-15 18:12:47.618609] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.123 qpair failed and we were unable to recover it. 00:26:54.123 [2024-07-15 18:12:47.618842] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.618872] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.619192] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.619222] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.619536] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.619566] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.619793] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.619808] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.620052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.620066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.620323] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.620339] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.620540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.620556] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.620858] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.620873] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.621125] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.621140] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.621380] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.621396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.621546] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.621561] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.621815] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.621830] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.622112] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.622143] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.622385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.622418] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.622602] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.622633] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.622913] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.622928] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.623075] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.623090] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.623400] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.623414] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.623714] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.623730] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.623933] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.623948] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.624130] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.624146] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.624419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.624436] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.624673] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.624689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.624967] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.624997] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.625317] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.625349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.625625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.625640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.625843] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.625858] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.626107] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.626122] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.626341] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.626356] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.626559] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.626574] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.626703] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.626718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.626957] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.626972] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.627236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.627252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.627529] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.627544] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.627695] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.627709] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.628005] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.628019] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.628220] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.628245] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.628499] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.628513] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.628707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.628721] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.628849] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.628863] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.629088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.629103] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.629397] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.629414] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.629561] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.629578] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.629798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.629815] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.630090] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.630104] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.630312] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.630329] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.630552] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.630567] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.630771] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.630786] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.630962] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.630976] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.631193] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.124 [2024-07-15 18:12:47.631207] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.124 qpair failed and we were unable to recover it. 00:26:54.124 [2024-07-15 18:12:47.631493] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.631509] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.631652] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.631666] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.631935] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.631949] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.632133] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.632147] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.632347] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.632363] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.632575] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.632590] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.632729] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.632744] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.632960] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.632974] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.633088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.633104] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.633248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.633264] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.633470] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.633484] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.633612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.633627] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.633818] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.633832] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.634119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.634134] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.634358] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.634372] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.634507] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.634522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.634696] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.634710] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.634903] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.634933] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.635117] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.635147] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.635414] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.635445] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.635726] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.635741] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.636040] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.636055] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.636394] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.636414] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.636563] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.636578] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.636771] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.636786] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.637038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.637052] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.637278] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.637293] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.637544] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.637559] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.637741] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.637756] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.637898] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.637912] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.638100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.638114] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.638310] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.638325] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.638649] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.638663] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.638979] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.638994] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.639271] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.639286] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.639536] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.639550] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.639753] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.639768] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.639992] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.640007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.640195] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.640210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.640451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.640466] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.640596] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.640611] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.640838] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.640852] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.641119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.641134] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.641257] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.641272] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.641554] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.641568] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.641751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.641766] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.642036] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.642050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.642313] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.642328] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.642442] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.642457] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.642726] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.642741] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.642889] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.642904] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.643179] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.643193] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.643449] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.643465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.643607] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.643622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.643875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.643889] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.644163] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.644177] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.644311] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.644325] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.125 [2024-07-15 18:12:47.644624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.125 [2024-07-15 18:12:47.644638] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.125 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.644903] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.644918] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.645181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.645196] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.645419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.645435] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.645715] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.645730] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.645875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.645889] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.646132] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.646177] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.646489] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.646526] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.646765] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.646798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.647033] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.647048] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.647270] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.647286] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.647509] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.647524] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.647708] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.647722] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.647904] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.647919] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.648210] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.648229] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.648447] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.648462] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.648720] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.648734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.648905] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.648920] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.649129] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.649144] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.649402] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.649422] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.649629] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.649643] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.649909] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.649923] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.650049] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.650063] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.650266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.650281] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.650512] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.650526] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.650713] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.650727] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.650933] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.650947] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.651100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.651115] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.651381] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.651396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.651599] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.651613] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.651888] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.651903] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.652155] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.652170] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.652421] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.652436] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.652561] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.652576] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.652719] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.652734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.652984] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.652998] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.653207] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.653222] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.653452] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.653467] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.653742] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.653756] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.653949] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.653964] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.654257] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.654271] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.654453] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.654468] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.654707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.654722] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.654973] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.654988] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.655139] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.655153] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.655280] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.655295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.655549] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.655567] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.655698] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.655712] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.655970] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.655984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.656165] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.656180] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.656302] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.656317] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.656505] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.656520] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.656642] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.656657] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.656872] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.656887] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.657025] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.126 [2024-07-15 18:12:47.657040] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.126 qpair failed and we were unable to recover it. 00:26:54.126 [2024-07-15 18:12:47.657223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.657371] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.657698] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.657728] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.657948] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.657963] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.658181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.658195] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.658392] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.658408] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.658660] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.658675] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.658923] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.658938] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.659194] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.659238] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.659472] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.659503] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.659781] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.659796] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.659922] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.659936] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.660207] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.660221] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.660353] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.660369] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.660557] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.660571] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.660706] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.660721] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.660824] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.660839] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.660977] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.660991] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.661113] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.661127] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.661375] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.661393] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.661578] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.661592] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.661771] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.661785] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.661896] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.661910] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.662172] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.662186] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.662368] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.662383] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.662582] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.662597] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.662783] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.662797] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.662989] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.663003] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.663193] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.663207] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.663404] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.663418] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.663525] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.663539] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.663822] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.663837] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.664040] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.664056] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.664318] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.664334] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.664476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.664492] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.664681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.664695] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.664816] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.664830] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.665057] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.665071] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.665155] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.665168] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.665375] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.665389] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.665593] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.665607] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.665784] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.665798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.665985] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.666000] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.666189] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.666219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.666563] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.666594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.666769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.666802] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.666982] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.667000] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.667141] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.667155] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.667334] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.667349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.667628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.667642] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.667790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.667803] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.667933] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.667947] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.668217] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.668244] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.127 qpair failed and we were unable to recover it. 00:26:54.127 [2024-07-15 18:12:47.668452] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.127 [2024-07-15 18:12:47.668466] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.668671] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.668685] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.668830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.668845] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.669115] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.669129] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.669323] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.669338] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.669447] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.669461] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.669730] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.669744] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.669942] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.669957] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.670141] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.670155] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.670264] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.670277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.670471] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.670485] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.670594] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.670608] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.670852] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.670865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.671040] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.671054] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.671236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.671251] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.671369] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.671383] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.671508] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.671522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.671633] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.671647] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.671822] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.671836] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.672102] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.672116] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.672323] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.672337] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.672486] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.672500] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.672676] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.672690] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.672938] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.672952] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.673156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.673170] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.673374] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.673388] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.673588] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.673602] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.673849] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.673863] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.674153] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.674167] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.674287] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.674302] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.674475] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.674489] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.674704] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.674718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.674971] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.674985] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.675198] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.675211] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.675514] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.675584] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.675862] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.675895] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.676126] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.676136] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.676315] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.676327] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.676564] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.676574] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.676777] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.676788] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.676968] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.676979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.677154] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.677164] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.677391] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.677403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.677584] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.677594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.677840] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.677870] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.678041] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.678070] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.678400] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.678431] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.678649] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.678688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.128 [2024-07-15 18:12:47.678965] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.128 [2024-07-15 18:12:47.678976] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.128 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.679241] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.679251] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.679486] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.679496] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.679753] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.679764] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.679962] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.679972] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.680279] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.680311] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.680550] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.680580] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.680854] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.680883] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.681040] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.681070] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.681393] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.681424] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.681723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.681753] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.681976] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.681987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.682218] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.682232] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.682447] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.682458] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.682627] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.682638] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.682887] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.682897] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.683156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.683166] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.683354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.683365] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.683554] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.683565] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.683759] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.683770] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.683977] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.683987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.684167] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.684177] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.684430] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.684441] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.684675] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.684685] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.684942] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.684953] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.685213] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.685223] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.685413] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.685430] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.685626] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.685640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.685883] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.685896] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.686163] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.686177] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.686368] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.686382] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.686638] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.686651] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.686911] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.686925] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.687116] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.687131] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.687342] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.687356] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.687622] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.687636] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.687887] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.687900] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.688140] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.688153] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.688329] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.688343] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.688538] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.688551] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.688745] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.688758] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.688950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.688965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.689182] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.689196] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.689435] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.689450] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.689739] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.689753] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.689996] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.690009] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.690249] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.690263] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.690389] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.129 [2024-07-15 18:12:47.690403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.129 qpair failed and we were unable to recover it. 00:26:54.129 [2024-07-15 18:12:47.690681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.690695] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.690982] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.690996] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.691269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.691283] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.691594] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.691624] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.691844] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.691874] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.692078] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.692113] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.692393] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.692423] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.692693] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.692724] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.693036] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.693049] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.693336] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.693350] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.693609] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.693623] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.693765] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.693779] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.693996] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.694010] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.694274] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.694288] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.694504] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.694519] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.694758] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.694772] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.695017] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.695031] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.695235] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.695250] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.695555] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.695585] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.695842] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.695872] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.696149] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.696180] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.696477] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.696509] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.696745] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.696775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.697047] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.697061] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.697300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.697314] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.697557] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.697571] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.697837] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.697850] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.698042] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.698055] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.698258] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.698272] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.698407] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.698421] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.698549] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.698562] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.698773] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.698787] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.698980] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.698997] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.699179] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.699193] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.699385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.699399] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.699628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.699657] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.699954] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.699983] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.700255] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.700287] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.700463] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.700492] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.700777] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.700791] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.701062] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.701076] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.701196] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.701210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.701410] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.701425] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.701717] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.701747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.701918] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.701948] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.702223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.702265] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.702567] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.702597] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.702889] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.130 [2024-07-15 18:12:47.702918] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.130 qpair failed and we were unable to recover it. 00:26:54.130 [2024-07-15 18:12:47.703144] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.703158] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.703424] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.703439] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.703631] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.703645] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.703916] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.703930] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.704108] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.704122] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.704418] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.704432] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.704672] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.704686] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.704950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.704964] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.705218] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.705237] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.705467] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.705481] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.705652] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.705666] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.705948] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.705964] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.706261] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.706275] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.706448] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.706462] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.706719] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.706733] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.707006] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.707020] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.707259] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.707274] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.707424] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.707437] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.707703] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.707717] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.708006] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.708019] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.708279] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.708293] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.708400] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.708414] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.708671] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.708685] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.708880] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.708894] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.709063] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.709077] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.709370] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.709409] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.709658] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.709689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.709964] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.709975] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.710267] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.710278] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.710465] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.710475] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.710609] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.710620] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.710853] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.710863] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.711089] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.711100] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.711349] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.711360] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.711567] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.711578] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.711812] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.711822] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.712083] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.712093] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.712375] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.712386] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.712648] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.712662] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.712897] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.712907] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.713102] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.713113] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.131 [2024-07-15 18:12:47.713372] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.131 [2024-07-15 18:12:47.713382] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.131 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.713663] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.713674] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.713936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.713946] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.714210] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.714221] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.714504] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.714514] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.714678] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.714689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.714866] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.714876] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.715056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.715067] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.715326] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.715336] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.715549] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.715559] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.715742] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.715754] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.715888] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.715898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.716104] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.716113] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.716350] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.716361] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.716534] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.716544] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.716731] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.716741] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.717003] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.717014] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.717187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.717197] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.717392] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.717404] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.717609] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.717619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.717749] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.717760] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.717960] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.717970] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.718234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.718245] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.718456] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.718466] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.718683] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.718699] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.718971] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.718985] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.719208] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.719222] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.719359] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.719373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.719688] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.719702] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.719990] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.720004] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.720208] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.720222] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.720493] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.720507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.720652] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.720666] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.720793] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.720807] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.721074] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.721088] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.721279] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.721293] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.721534] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.721548] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.721813] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.721827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.722088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.722102] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.722341] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.722355] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.722473] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.722487] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.722675] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.722689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.722859] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.722873] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.723040] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.723053] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.723326] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.723339] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.723559] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.723572] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.723813] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.723827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.724107] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.724121] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.724329] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.724344] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.724551] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.724565] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.724682] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.724696] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.724868] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.724885] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.725055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.725069] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.725308] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.725321] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.725591] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.725605] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.725794] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.725808] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.726067] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.726081] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.726381] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.132 [2024-07-15 18:12:47.726394] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.132 qpair failed and we were unable to recover it. 00:26:54.132 [2024-07-15 18:12:47.726608] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.726622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.726792] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.726806] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.727008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.727022] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.727268] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.727282] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.727473] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.727487] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.727673] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.727687] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.727950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.727964] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.728171] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.728186] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.728402] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.728417] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.728594] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.728608] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.728799] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.728813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.729053] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.729102] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.729392] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.729424] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.729662] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.729676] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.729869] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.729883] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.730090] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.730114] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.730329] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.730343] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.730603] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.730617] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.730799] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.730813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.731124] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.731153] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.731457] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.731493] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.731707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.731720] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.732007] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.732021] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.732406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.732423] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.732654] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.732670] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.732804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.732818] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.733087] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.733102] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.733214] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.733234] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.733451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.733465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.733650] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.733663] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.733901] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.733915] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.734175] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.734189] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.734368] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.734382] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.734596] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.734610] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.734735] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.734748] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.734964] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.734978] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.735103] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.735117] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.735221] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.735240] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.735436] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.735450] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.735718] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.735749] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.735975] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.736005] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.736214] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.736261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.736507] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.736537] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.736830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.736860] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.737158] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.737171] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.737479] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.737494] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.737803] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.737816] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.738131] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.738167] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.738339] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.738370] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.738671] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.738700] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.738946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.738960] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.133 [2024-07-15 18:12:47.739142] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.133 [2024-07-15 18:12:47.739155] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.133 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.739393] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.739408] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.739602] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.739616] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.739799] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.739813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.740075] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.740088] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.740347] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.740361] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.740621] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.740635] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.740888] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.740902] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.741188] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.741201] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.741465] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.741479] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.741708] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.741722] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.741991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.742005] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.742241] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.742256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.742504] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.742518] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.742785] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.742798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.742993] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.743007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.743292] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.743307] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.743495] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.743508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.743694] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.743708] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.743997] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.744011] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.744255] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.744269] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.744454] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.744468] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.744731] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.744745] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.744944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.744958] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.745245] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.745260] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.745376] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.745389] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.745649] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.745663] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.745906] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.745920] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.746189] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.746203] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.746469] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.746484] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.746723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.746737] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.746999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.747013] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.747201] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.747214] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.747410] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.747423] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.747673] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.747687] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.747952] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.747965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.748245] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.748260] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.748556] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.748593] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.748803] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.748818] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.748998] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.749012] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.749273] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.749289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.749549] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.749579] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.749805] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.749835] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.750002] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.750032] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.750323] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.750339] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.750584] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.750598] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.750837] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.134 [2024-07-15 18:12:47.750851] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.134 qpair failed and we were unable to recover it. 00:26:54.134 [2024-07-15 18:12:47.751041] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.751055] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.751260] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.751291] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.751579] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.751610] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.751892] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.751910] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.752090] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.752104] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.752297] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.752313] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.752506] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.752519] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.752701] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.752715] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.752975] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.753015] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.753249] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.753279] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.753592] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.753606] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.753891] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.753904] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.754095] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.754108] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.754382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.754413] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.754704] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.754733] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.754956] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.754986] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.755286] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.755318] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.755627] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.755657] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.755933] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.755962] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.756256] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.756271] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.756537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.756551] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.756817] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.756830] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.757098] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.757112] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.757378] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.757393] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.757569] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.757582] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.757773] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.757787] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.758029] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.758042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.758312] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.758326] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.758535] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.758549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.758814] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.758827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.759041] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.759079] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.759381] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.759398] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.759690] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.759705] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.759954] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.759968] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.760205] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.760219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.760466] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.760480] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.760605] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.760619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.760875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.760888] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.760999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.761013] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.761259] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.761274] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.761541] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.761555] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.761815] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.135 [2024-07-15 18:12:47.761829] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.135 qpair failed and we were unable to recover it. 00:26:54.135 [2024-07-15 18:12:47.762111] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.762125] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.762365] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.762379] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.762577] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.762591] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.762853] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.762883] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.763167] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.763197] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.763454] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.763485] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.763732] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.763762] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.764061] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.764075] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.764319] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.764333] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.764582] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.764596] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.764791] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.764804] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.765064] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.765078] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.765283] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.765298] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.765563] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.765577] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.765834] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.765848] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.765988] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.766004] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.766217] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.766256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.766477] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.766508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.766723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.766753] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.767062] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.767093] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.767383] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.767415] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.767641] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.767671] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.767985] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.768015] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.768321] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.768352] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.768567] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.768597] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.768778] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.768792] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.769053] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.769083] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.769381] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.769413] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.769659] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.769689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.769965] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.769979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.770234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.770248] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.770453] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.770467] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.770754] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.770767] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.770887] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.770900] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.771092] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.771124] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.771424] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.771454] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.771752] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.771781] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.772026] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.772063] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.772240] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.772255] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.772540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.772554] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.772815] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.772828] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.773091] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.773104] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.773385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.773404] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.773573] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.773587] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.773883] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.773897] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.774134] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.774147] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.774400] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.774415] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.774615] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.774628] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.774795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.774808] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.775096] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.775125] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.775443] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.775474] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.775702] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.775731] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.136 qpair failed and we were unable to recover it. 00:26:54.136 [2024-07-15 18:12:47.775973] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.136 [2024-07-15 18:12:47.776002] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.776306] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.776339] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.776634] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.776664] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.776960] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.776974] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.777149] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.777163] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.777332] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.777346] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.777530] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.777543] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.777730] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.777744] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.777994] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.778024] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.778260] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.778290] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.778580] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.778610] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.778823] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.778853] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.779057] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.779086] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.779281] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.779303] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.779493] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.779506] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.779768] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.779798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.780092] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.780122] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.780426] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.780440] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.780566] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.780580] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.780862] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.780891] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.781172] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.781201] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.781510] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.781541] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.781745] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.781774] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.782009] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.782039] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.782268] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.782300] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.782588] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.782617] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.782888] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.782917] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.783239] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.783253] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.783540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.783553] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.783746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.783760] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.783990] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.784004] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.784265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.784279] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.784540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.784553] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.784838] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.784869] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.785092] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.785121] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.785341] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.785373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.785624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.785654] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.785944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.785975] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.786198] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.786212] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.786511] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.786547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.786790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.786826] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.787103] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.787118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.787413] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.787430] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.787561] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.787575] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.787838] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.787852] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.788049] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.788064] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.788254] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.788268] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.788450] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.788463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.788728] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.788757] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.789073] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.789102] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.789400] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.789414] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.789554] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.789567] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.789831] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.789845] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.790101] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.790115] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.790320] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.790334] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.137 [2024-07-15 18:12:47.790602] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.137 [2024-07-15 18:12:47.790632] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.137 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.790842] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.790872] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.791077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.791107] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.791344] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.791376] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.791681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.791710] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.791918] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.791949] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.792196] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.792210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.792483] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.792498] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.792683] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.792696] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.792960] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.792974] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.793183] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.793197] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.793475] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.793490] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.793728] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.793742] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.794004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.794017] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.794252] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.794267] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.794526] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.794540] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.794673] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.794692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.794947] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.794978] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.795266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.795297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.795518] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.795547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.795760] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.795790] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.796066] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.796096] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.796299] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.796330] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.796602] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.796631] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.796951] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.796980] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.797176] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.797190] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.797376] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.797390] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.797671] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.797701] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.797995] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.798025] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.798325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.798356] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.798655] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.798685] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.798898] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.798911] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.799172] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.799185] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.799445] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.799459] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.799714] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.799728] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.799985] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.799999] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.800252] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.800266] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.800506] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.800520] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.800729] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.800743] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.801069] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.801083] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.801342] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.801357] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.138 [2024-07-15 18:12:47.801615] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.138 [2024-07-15 18:12:47.801630] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.138 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.801865] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.801879] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.802121] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.802135] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.802331] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.802345] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.802552] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.802567] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.802734] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.802748] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.802873] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.802888] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.803083] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.803097] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.803188] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.803201] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.803487] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.803501] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.803709] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.803723] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.803985] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.804014] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.804310] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.804341] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.804556] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.804586] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.804829] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.804858] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.805080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.805096] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.805349] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.805363] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.805627] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.805641] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.805831] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.805845] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.806092] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.806106] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.806390] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.806405] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.806584] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.806598] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.806831] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.806844] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.806976] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.806990] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.807171] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.807185] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.807422] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.807438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.807640] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.807654] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.807766] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.807780] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.808038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.808052] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.808307] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.808338] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.808630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.808660] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.808942] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.808972] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.809187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.809216] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.809460] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.809490] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.809784] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.809814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.810022] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.810052] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.810266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.810298] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.810518] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.810547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.810771] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.810819] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.811080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.811093] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.811376] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.811390] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.811498] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.811512] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.811705] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.811719] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.811884] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.811897] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.812130] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.812144] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.812383] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.812398] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.812686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.139 [2024-07-15 18:12:47.812701] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.139 qpair failed and we were unable to recover it. 00:26:54.139 [2024-07-15 18:12:47.812977] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.812991] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.813236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.813250] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.813501] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.813515] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.813708] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.813722] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.814038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.814053] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.814317] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.814332] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.814517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.814531] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.814791] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.814805] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.815008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.815025] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.815302] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.815316] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.815581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.815594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.815893] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.815934] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.816151] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.816181] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.816433] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.816464] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.816687] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.816718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.817009] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.817039] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.817278] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.817292] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.817468] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.817482] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.817751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.817780] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.818033] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.818062] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.818285] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.818299] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.818540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.818553] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.818821] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.818836] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.819089] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.819103] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.819280] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.819294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.819493] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.819507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.819773] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.819802] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.820059] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.820088] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.820289] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.820303] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.820543] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.820572] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.820877] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.820907] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.821143] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.821156] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.821359] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.821373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.821581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.821595] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.821784] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.821814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.822111] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.822141] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.822439] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.822470] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.822711] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.822741] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.822971] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.823002] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.823270] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.823284] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.823479] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.823493] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.823673] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.823688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.823906] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.823937] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.824158] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.824188] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.824402] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.824434] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.824658] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.824688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.824998] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.825012] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.140 [2024-07-15 18:12:47.825301] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.140 [2024-07-15 18:12:47.825315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.140 qpair failed and we were unable to recover it. 00:26:54.141 [2024-07-15 18:12:47.825436] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.141 [2024-07-15 18:12:47.825452] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.141 qpair failed and we were unable to recover it. 00:26:54.141 [2024-07-15 18:12:47.825582] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.141 [2024-07-15 18:12:47.825596] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.141 qpair failed and we were unable to recover it. 00:26:54.141 [2024-07-15 18:12:47.825777] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.141 [2024-07-15 18:12:47.825791] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.141 qpair failed and we were unable to recover it. 00:26:54.141 [2024-07-15 18:12:47.825908] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.141 [2024-07-15 18:12:47.825921] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.141 qpair failed and we were unable to recover it. 00:26:54.141 [2024-07-15 18:12:47.826118] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.141 [2024-07-15 18:12:47.826148] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.141 qpair failed and we were unable to recover it. 00:26:54.141 [2024-07-15 18:12:47.826382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.141 [2024-07-15 18:12:47.826414] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.141 qpair failed and we were unable to recover it. 00:26:54.141 [2024-07-15 18:12:47.826664] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.141 [2024-07-15 18:12:47.826694] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.141 qpair failed and we were unable to recover it. 00:26:54.141 [2024-07-15 18:12:47.826985] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.141 [2024-07-15 18:12:47.826999] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.141 qpair failed and we were unable to recover it. 00:26:54.141 [2024-07-15 18:12:47.827197] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.141 [2024-07-15 18:12:47.827210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.141 qpair failed and we were unable to recover it. 00:26:54.141 [2024-07-15 18:12:47.827425] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.141 [2024-07-15 18:12:47.827440] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.141 qpair failed and we were unable to recover it. 00:26:54.141 [2024-07-15 18:12:47.827631] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.141 [2024-07-15 18:12:47.827645] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.141 qpair failed and we were unable to recover it. 00:26:54.141 [2024-07-15 18:12:47.827898] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.141 [2024-07-15 18:12:47.827911] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.141 qpair failed and we were unable to recover it. 00:26:54.141 [2024-07-15 18:12:47.828221] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.141 [2024-07-15 18:12:47.828239] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.141 qpair failed and we were unable to recover it. 00:26:54.141 [2024-07-15 18:12:47.828446] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.141 [2024-07-15 18:12:47.828460] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.141 qpair failed and we were unable to recover it. 00:26:54.141 [2024-07-15 18:12:47.828573] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.141 [2024-07-15 18:12:47.828587] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.141 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.828772] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.828787] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.829065] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.829079] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.829365] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.829380] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.829631] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.829644] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.829757] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.829771] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.829955] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.829969] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.830245] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.830275] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.830553] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.830582] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.830787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.830817] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.831097] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.831110] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.831376] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.831390] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.831687] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.831700] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.831930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.831959] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.832151] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.832162] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.832399] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.832411] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.832694] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.832705] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.832977] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.832987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.833171] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.833181] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.833456] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.833466] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.833727] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.833737] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.834006] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.834017] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.834125] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.834135] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.834338] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.834348] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.834468] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.834478] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.834657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.834667] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.834912] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.834926] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.835158] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.835168] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.835354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.835364] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.835626] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.835654] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.835875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.422 [2024-07-15 18:12:47.835905] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.422 qpair failed and we were unable to recover it. 00:26:54.422 [2024-07-15 18:12:47.836194] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.836223] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.836532] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.836562] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.836768] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.836798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.837068] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.837106] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.837210] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.837220] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.837402] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.837412] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.837663] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.837692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.837924] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.837954] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.838222] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.838262] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.838490] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.838520] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.838686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.838715] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.838940] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.838969] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.839260] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.839291] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.839448] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.839478] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.839764] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.839793] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.840037] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.840066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.840337] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.840367] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.840584] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.840614] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.840905] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.840934] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.841126] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.841135] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.841309] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.841319] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.841530] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.841560] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.841791] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.841825] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.842124] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.842154] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.842363] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.842376] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.842561] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.842576] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.842696] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.842709] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.842835] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.842848] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.843027] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.843040] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.843228] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.843242] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.843362] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.843376] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.843555] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.843568] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.843745] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.843759] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.843928] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.843942] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.844072] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.844086] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.844254] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.844270] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.844507] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.844521] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.844793] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.844823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.845155] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.845185] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.845390] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.845404] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.845556] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.845570] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.845808] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.845821] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.846000] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.846013] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.846255] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.846286] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.846445] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.846475] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.846637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.846667] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.846967] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.846996] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.847221] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.847261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.847475] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.847489] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.847739] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.847753] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.847857] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.847871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.848042] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.848055] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.848238] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.848252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.848457] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.848470] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.848644] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.848658] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.423 [2024-07-15 18:12:47.848837] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.423 [2024-07-15 18:12:47.848866] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.423 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.849004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.849034] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.849323] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.849354] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.849523] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.849553] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.849778] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.849807] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.850076] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.850105] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.850315] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.850329] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.850526] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.850554] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.850751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.850762] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.850882] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.850893] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.851092] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.851120] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.851342] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.851374] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.851603] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.851633] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.851762] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.851791] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.851949] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.851985] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.852089] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.852099] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.852279] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.852290] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.852541] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.852550] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.852681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.852691] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.852891] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.852901] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.853091] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.853104] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.853220] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.853234] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.853395] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.853405] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.853526] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.853536] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.853714] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.853724] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.853889] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.853898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.854075] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.854086] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.854176] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.854185] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.854299] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.854309] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.854480] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.854490] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.854672] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.854682] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.854845] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.854855] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.854966] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.854976] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.855155] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.855164] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.855329] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.855340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.855448] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.855458] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.855637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.855647] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.855766] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.855776] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.855946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.855956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.856056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.856066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.856187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.856197] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.856388] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.856399] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.856515] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.856525] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.856707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.856716] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.856821] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.856831] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.856942] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.856952] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.857151] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.857160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.857285] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.857296] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.857416] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.857426] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.857528] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.857538] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.857730] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.857740] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.857915] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.857925] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.858096] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.858106] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.858272] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.858283] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.858564] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.858594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.858799] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.858828] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.424 qpair failed and we were unable to recover it. 00:26:54.424 [2024-07-15 18:12:47.859051] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.424 [2024-07-15 18:12:47.859080] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.859380] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.859390] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.859487] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.859496] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.859670] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.859680] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.859837] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.859849] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.860023] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.860033] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.860248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.860259] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.860511] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.860521] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.860685] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.860696] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.860805] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.860814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.860924] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.860934] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.861055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.861065] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.861318] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.861328] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.861581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.861591] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.861709] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.861719] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.861890] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.861900] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.862070] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.862080] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.862269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.862299] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.862549] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.862579] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.862781] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.862811] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.863026] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.863055] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.863268] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.863278] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.863379] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.863388] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.863559] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.863569] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.863823] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.863852] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.864122] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.864151] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.864370] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.864380] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.864565] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.864594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.864747] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.864776] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.864927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.864957] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.865173] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.865193] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.865453] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.865463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.865576] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.865586] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.865813] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.865823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.865956] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.865985] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.866131] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.866160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.866384] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.866415] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.866614] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.866623] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.866806] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.866836] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.867081] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.867110] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.867381] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.867412] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.867682] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.867711] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.867956] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.867986] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.868208] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.868250] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.868396] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.868408] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.868583] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.868619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.868836] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.868865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.869093] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.869103] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.425 qpair failed and we were unable to recover it. 00:26:54.425 [2024-07-15 18:12:47.869282] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.425 [2024-07-15 18:12:47.869292] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.869509] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.869519] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.869784] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.869795] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.869968] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.869978] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.870160] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.870201] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.870515] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.870546] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.870831] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.870865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.871026] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.871036] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.871320] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.871351] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.871524] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.871552] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.871740] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.871771] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.872022] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.872052] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.872342] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.872373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.872593] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.872622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.872841] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.872871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.873148] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.873178] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.873422] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.873452] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.873611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.873641] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.873859] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.873888] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.874040] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.874079] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.874203] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.874212] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.874415] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.874425] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.874598] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.874628] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.874857] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.874888] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.875098] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.875127] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.875339] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.875350] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.875466] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.875476] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.875654] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.875664] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.875776] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.875786] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.875939] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.875949] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.876055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.876064] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.876253] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.876263] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.876382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.876392] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.876578] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.876588] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.876827] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.876837] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.877098] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.877127] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.877315] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.877351] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.877558] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.877588] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.877749] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.877777] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.878005] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.878034] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.878269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.878299] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.878519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.878548] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.878773] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.878801] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.879007] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.879035] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.879234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.879244] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.879427] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.879456] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.879598] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.879626] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.879776] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.879807] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.880012] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.880041] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.880260] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.880290] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.880447] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.880476] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.880694] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.880724] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.880878] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.880907] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.881056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.881084] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.881299] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.881329] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.881608] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.881618] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.881804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.881814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.426 qpair failed and we were unable to recover it. 00:26:54.426 [2024-07-15 18:12:47.882008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.426 [2024-07-15 18:12:47.882018] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.882270] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.882280] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.882385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.882395] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.882503] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.882513] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.882690] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.882699] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.882807] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.882817] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.882990] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.883001] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.883277] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.883288] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.883402] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.883412] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.883593] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.883603] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.883755] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.883764] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.884001] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.884011] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.884239] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.884249] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.884354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.884363] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.884594] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.884604] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.884738] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.884747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.884891] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.884901] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.885153] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.885162] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.885265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.885274] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.885435] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.885444] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.885549] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.885559] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.885662] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.885672] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.885763] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.885772] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.885887] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.885897] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.886032] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.886042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.886223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.886236] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.886397] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.886406] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.886511] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.886520] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.886646] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.886656] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.886776] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.886785] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.886956] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.886966] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.887135] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.887144] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.887366] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.887377] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.887549] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.887559] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.887668] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.887686] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.887781] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.887790] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.887907] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.887917] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.888024] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.888033] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.888216] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.888230] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.888353] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.888369] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.888548] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.888558] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.888733] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.888743] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.888909] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.888918] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.889038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.889048] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.889160] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.889170] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.889342] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.889352] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.889581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.889593] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.889772] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.889781] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.890008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.890018] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.890133] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.890143] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.890320] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.890331] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.890608] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.890618] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.890781] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.890807] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.890918] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.890928] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.891078] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.891088] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.891223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.891238] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.891345] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.427 [2024-07-15 18:12:47.891354] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.427 qpair failed and we were unable to recover it. 00:26:54.427 [2024-07-15 18:12:47.891535] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.891546] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.891772] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.891801] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.891942] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.891970] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.892119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.892148] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.892373] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.892383] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.892616] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.892625] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.892870] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.892879] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.893052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.893062] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.893319] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.893328] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.893510] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.893520] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.893722] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.893752] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.893956] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.893985] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.894140] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.894169] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.894345] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.894354] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.894463] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.894472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.894639] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.894650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.894798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.894807] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.894985] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.894995] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.895185] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.895195] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.895300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.895309] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.895469] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.895479] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.895603] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.895612] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.895774] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.895783] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.895955] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.895966] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.896124] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.896133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.896255] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.896265] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.896343] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.896351] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.896462] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.896472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.896578] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.896588] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.896709] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.896721] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.896856] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.896865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.896986] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.896996] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.897229] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.897240] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.897352] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.897361] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.897454] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.897465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.897570] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.897579] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.897671] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.897680] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.897769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.897778] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.898007] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.898017] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.898177] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.898187] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.898294] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.898304] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.898479] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.898490] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.898605] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.898615] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.898860] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.898870] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.899040] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.899050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.899156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.428 [2024-07-15 18:12:47.899165] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.428 qpair failed and we were unable to recover it. 00:26:54.428 [2024-07-15 18:12:47.899278] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.899288] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.899391] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.899400] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.899509] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.899519] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.899702] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.899711] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.899830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.899840] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.900010] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.900020] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.900164] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.900173] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.900352] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.900363] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.900546] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.900556] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.900734] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.900743] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.900825] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.900834] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.900928] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.900939] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.901182] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.901193] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.901302] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.901312] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.901492] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.901501] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.901754] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.901764] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.901926] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.901936] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.902087] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.902097] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.902212] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.902221] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.902456] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.902465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.902566] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.902575] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.902671] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.902680] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.902804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.902814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.903068] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.903079] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.903244] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.903254] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.903431] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.903441] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.903547] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.903557] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.903796] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.903806] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.903972] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.903981] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.904154] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.904164] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.904287] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.904297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.904399] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.904409] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.904519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.904529] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.904642] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.904651] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.904718] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.904727] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.904817] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.904826] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.904993] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.905003] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.905078] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.905087] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.905259] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.905269] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.905425] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.905435] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.905551] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.905561] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.905739] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.905749] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.905946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.905956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.906065] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.906074] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.906243] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.906253] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.906490] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.906500] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.906661] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.906670] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.906777] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.906787] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.906913] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.906922] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.907033] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.907042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.907214] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.907234] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.429 [2024-07-15 18:12:47.907339] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.429 [2024-07-15 18:12:47.907349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.429 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.907449] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.907458] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.907551] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.907560] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.907738] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.907747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.907862] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.907872] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.908039] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.908049] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.908147] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.908157] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.908395] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.908405] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.908562] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.908572] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.908706] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.908715] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.908936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.908945] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.909122] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.909132] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.909228] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.909240] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.909547] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.909557] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.909647] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.909656] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.909857] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.909887] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.910036] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.910065] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.910292] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.910322] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.910520] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.910529] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.910640] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.910649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.910825] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.910835] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.910947] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.910957] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.911147] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.911157] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.911360] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.911369] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.911487] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.911497] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.911630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.911640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.911814] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.911824] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.911950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.911960] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.912064] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.912073] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.912180] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.912191] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.912381] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.912390] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.912572] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.912582] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.912753] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.912763] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.912921] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.912930] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.913100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.913110] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.913241] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.913251] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.913478] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.913488] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.913656] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.913666] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.913774] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.913784] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.913953] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.913962] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.914128] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.914139] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.914308] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.914318] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.914490] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.914500] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.914670] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.914679] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.914788] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.914798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.915056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.915065] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.915163] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.915175] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.915294] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.915304] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.430 qpair failed and we were unable to recover it. 00:26:54.430 [2024-07-15 18:12:47.915415] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.430 [2024-07-15 18:12:47.915425] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.915655] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.915665] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.915840] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.915849] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.915951] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.915961] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.916058] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.916070] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.916235] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.916246] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.916475] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.916484] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.916594] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.916604] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.916764] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.916775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.917039] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.917049] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.917229] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.917239] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.917423] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.917432] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.917540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.917549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.917732] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.917743] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.917860] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.917871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.917970] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.917980] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.918152] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.918161] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.918338] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.918349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.918578] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.918588] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.918757] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.918767] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.918892] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.918902] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.919063] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.919073] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.919234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.919245] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.919497] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.919526] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.919682] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.919711] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.919921] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.919949] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.920210] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.920220] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.920337] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.920347] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.920518] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.920528] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.920618] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.920627] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.920837] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.920847] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.921085] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.921115] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.921322] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.921352] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.921498] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.921527] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.921789] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.921799] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.922048] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.922058] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.922178] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.922187] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.922347] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.922358] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.922534] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.922543] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.922726] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.922755] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.922957] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.922987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.923217] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.923256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.923406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.923435] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.923627] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.923637] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.923823] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.923857] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.924076] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.924104] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.924309] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.924319] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.924498] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.924508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.431 [2024-07-15 18:12:47.924610] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.431 [2024-07-15 18:12:47.924619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.431 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.924846] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.924856] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.925032] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.925042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.925236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.925246] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.925410] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.925420] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.925488] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.925497] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.925628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.925636] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.925825] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.925835] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.925953] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.925964] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.926170] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.926179] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.926446] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.926477] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.926715] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.926745] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.926956] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.926986] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.927148] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.927158] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.927336] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.927347] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.927461] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.927472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.927589] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.927599] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.927825] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.927835] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.927997] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.928008] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.928185] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.928195] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.928419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.928449] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.928691] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.928721] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.928873] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.928902] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.929051] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.929092] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.929262] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.929273] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.929461] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.929471] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.929633] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.929643] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.929823] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.929852] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.929991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.930019] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.930261] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.930292] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.930496] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.930526] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.930741] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.930772] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.930925] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.930954] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.931222] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.931237] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.931419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.931431] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.931545] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.931554] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.931718] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.931729] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.931890] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.931900] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.932113] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.932142] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.932297] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.932327] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.932597] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.932627] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.932848] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.932878] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.933040] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.933069] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.933276] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.933286] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.933461] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.933471] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.933658] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.933668] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.933871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.933881] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.934050] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.432 [2024-07-15 18:12:47.934060] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.432 qpair failed and we were unable to recover it. 00:26:54.432 [2024-07-15 18:12:47.934186] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.934195] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.934357] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.934367] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.934493] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.934502] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.934598] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.934610] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.934787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.934797] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.935001] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.935011] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.935188] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.935198] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.935384] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.935394] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.935558] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.935569] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.935732] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.935742] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.935854] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.935864] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.935983] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.935993] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.936097] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.936107] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.936266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.936278] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.936395] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.936407] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.936517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.936527] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.936778] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.936788] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.936879] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.936888] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.937084] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.937095] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.937187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.937198] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.937313] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.937323] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.937430] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.937441] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.937617] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.937627] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.937832] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.937842] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.937997] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.938007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.938106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.938116] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.938217] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.938234] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.938406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.938415] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.938596] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.938608] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.938716] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.938726] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.938824] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.938833] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.938952] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.938962] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.939063] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.939072] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.939237] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.939247] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.939412] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.939422] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.939526] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.939536] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.939626] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.939635] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.939744] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.939754] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.940014] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.940024] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.940223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.940237] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.940321] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.940330] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.940420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.940436] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.940746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.940756] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.940861] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.940871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.940991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.941002] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.941096] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.941105] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.941191] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.941203] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.941306] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.941317] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.941427] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.941436] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.941504] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.941513] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.941623] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.941632] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.941749] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.941760] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.942011] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.942021] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.942119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.942129] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.942355] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.942366] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.433 [2024-07-15 18:12:47.942596] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.433 [2024-07-15 18:12:47.942607] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.433 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.942723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.942732] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.942889] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.942899] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.943001] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.943010] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.943109] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.943119] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.943270] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.943280] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.943390] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.943400] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.943619] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.943630] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.943807] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.943816] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.943986] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.943995] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.944106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.944116] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.944220] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.944234] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.944325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.944335] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.944449] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.944460] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.944555] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.944564] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.944673] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.944683] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.944854] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.944865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.945036] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.945045] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.945304] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.945314] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.945413] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.945423] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.945524] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.945532] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.945702] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.945713] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.945814] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.945824] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.945946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.945955] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.946127] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.946136] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.946310] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.946320] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.946484] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.946494] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.946594] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.946603] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.946697] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.946707] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.946794] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.946803] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.947047] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.947058] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.947221] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.947235] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.947402] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.947413] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.947596] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.947606] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.947783] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.947793] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.947906] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.947916] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.948076] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.948086] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.948192] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.948202] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.948309] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.948318] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.948481] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.948490] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.948590] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.948599] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.948708] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.948718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.948833] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.948843] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.948945] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.948954] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.949123] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.949133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.949304] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.949315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.949515] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.949524] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.949728] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.949738] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.949897] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.949906] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.949999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.950009] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.950113] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.950123] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.950289] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.434 [2024-07-15 18:12:47.950300] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.434 qpair failed and we were unable to recover it. 00:26:54.434 [2024-07-15 18:12:47.950407] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.950417] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.950575] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.950587] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.950692] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.950701] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.950868] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.950878] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.950990] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.950999] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.951176] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.951185] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.951413] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.951424] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.951558] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.951568] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.951680] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.951689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.951766] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.951774] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.951941] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.951950] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.952115] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.952126] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.952283] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.952292] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.952384] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.952394] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.952499] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.952508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.952603] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.952614] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.952702] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.952711] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.952819] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.952830] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.952940] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.952950] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.953038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.953047] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.953137] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.953147] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.953257] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.953266] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.953342] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.953351] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.953448] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.953457] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.953702] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.953712] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.953820] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.953830] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.953996] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.954006] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.954170] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.954180] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.954297] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.954307] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.954403] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.954412] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.954576] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.954587] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.954821] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.954850] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.955062] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.955091] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.955314] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.955324] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.955498] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.955535] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.955673] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.955701] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.955851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.955881] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.956037] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.956066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.956233] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.956263] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.956469] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.956498] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.956698] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.956727] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.435 qpair failed and we were unable to recover it. 00:26:54.435 [2024-07-15 18:12:47.956994] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.435 [2024-07-15 18:12:47.957029] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.957240] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.957270] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.957426] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.957455] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.957675] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.957704] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.957909] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.957938] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.958175] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.958205] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.958437] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.958468] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.958613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.958641] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.958795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.958823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.959021] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.959051] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.959255] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.959286] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.959483] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.959493] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.959607] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.959616] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.959831] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.959860] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.960017] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.960046] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.960251] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.960282] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.960497] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.960526] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.960741] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.960770] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.961007] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.961037] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.961175] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.961185] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.961368] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.961395] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.961596] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.961625] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.961866] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.961895] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.962166] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.962195] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.962405] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.962435] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.962648] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.962658] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.962780] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.962790] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.962962] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.962971] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.963088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.963099] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.963264] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.963274] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.963381] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.963391] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.963554] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.963564] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.963732] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.963742] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.963995] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.964004] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.964118] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.964128] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.964289] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.964299] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.964503] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.964513] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.964617] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.964627] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.964694] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.964703] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.964813] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.964822] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.964932] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.964944] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.965128] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.965138] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.965295] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.965305] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.965413] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.965423] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.965521] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.965531] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.965641] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.965654] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.965901] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.965910] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.436 [2024-07-15 18:12:47.966017] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.436 [2024-07-15 18:12:47.966027] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.436 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.966197] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.966207] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.966382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.966392] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.966542] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.966552] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.966663] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.966673] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.966817] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.966826] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.967005] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.967015] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.967208] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.967249] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.967472] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.967501] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.967668] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.967696] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.967913] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.967941] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.968162] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.968190] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.968402] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.968411] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.968645] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.968674] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.968808] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.968837] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.969059] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.969088] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.969350] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.969360] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.969562] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.969571] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.969681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.969692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.969853] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.969863] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.969965] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.969976] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.970145] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.970155] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.970351] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.970361] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.970528] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.970538] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.970726] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.970735] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.970836] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.970846] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.971018] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.971028] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.971134] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.971145] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.971266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.971276] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.971382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.971392] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.971556] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.971566] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.971739] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.971748] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.971908] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.971919] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.972075] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.972085] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.972263] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.972273] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.972448] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.972458] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.972552] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.972561] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.972679] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.972689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.972856] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.972865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.972968] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.972978] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.973148] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.973185] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.973399] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.973429] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.973666] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.973695] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.973924] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.973953] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.974197] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.974242] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.974400] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.974409] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.974502] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.974511] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.974774] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.974784] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.974905] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.974915] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.975168] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.975178] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.975349] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.975359] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.975610] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.975639] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.975863] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.437 [2024-07-15 18:12:47.975893] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.437 qpair failed and we were unable to recover it. 00:26:54.437 [2024-07-15 18:12:47.976117] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.976147] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.976409] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.976418] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.976537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.976547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.976707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.976717] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.976961] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.976971] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.977080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.977089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.977284] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.977294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.977471] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.977483] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.977592] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.977602] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.977769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.977779] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.977902] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.977912] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.978139] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.978149] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.978251] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.978260] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.978443] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.978453] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.978618] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.978628] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.978825] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.978854] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.979072] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.979101] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.979302] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.979332] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.979556] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.979566] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.979751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.979760] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.980016] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.980046] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.980270] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.980301] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.980576] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.980606] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.980836] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.980865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.981070] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.981099] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.981305] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.981315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.981554] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.981582] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.981897] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.981927] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.982148] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.982177] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.982420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.982451] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.982720] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.982750] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.982970] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.982980] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.983146] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.983155] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.983384] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.983414] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.983560] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.983589] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.983819] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.983848] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.983996] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.984025] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.984291] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.984320] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.984611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.984639] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.984857] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.984886] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.985046] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.985075] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.985264] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.985294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.985565] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.985594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.985804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.985833] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.986035] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.986065] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.986350] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.986360] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.986517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.986527] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.986696] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.986733] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.986937] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.986967] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.987185] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.987214] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.987432] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.438 [2024-07-15 18:12:47.987441] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.438 qpair failed and we were unable to recover it. 00:26:54.438 [2024-07-15 18:12:47.987678] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.987706] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.987851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.987880] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.988029] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.988058] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.988258] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.988288] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.988440] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.988469] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.988708] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.988736] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.988885] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.988914] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.989155] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.989184] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.989408] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.989438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.989637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.989647] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.989825] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.989854] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.990123] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.990152] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.990375] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.990405] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.990615] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.990644] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.990960] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.990989] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.991237] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.991268] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.991561] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.991592] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.991751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.991780] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.991979] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.992008] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.992243] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.992290] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.992460] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.992490] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.992800] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.992828] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.993038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.993067] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.993240] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.993271] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.993434] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.993463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.993599] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.993627] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.993835] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.993845] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.994019] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.994048] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.994359] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.994391] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.994591] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.994621] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.994771] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.994799] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.994957] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.994984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.995253] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.995283] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.995506] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.995535] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.995660] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.995669] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.995905] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.995935] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.996147] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.996186] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.996346] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.996375] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.996660] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.996670] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.996845] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.996854] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.996973] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.439 [2024-07-15 18:12:47.996982] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.439 qpair failed and we were unable to recover it. 00:26:54.439 [2024-07-15 18:12:47.997159] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:47.997170] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:47.997334] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:47.997344] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:47.997445] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:47.997455] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:47.997646] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:47.997656] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:47.997839] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:47.997868] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:47.998089] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:47.998119] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:47.998432] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:47.998462] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:47.998687] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:47.998714] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:47.998942] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:47.998953] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:47.999129] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:47.999138] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:47.999325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:47.999336] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:47.999454] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:47.999465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:47.999643] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:47.999654] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:47.999842] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:47.999851] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.000018] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.000043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.000264] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.000295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.000617] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.000645] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.000905] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.000915] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.001009] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.001018] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.001177] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.001186] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.001364] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.001374] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.001488] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.001498] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.001727] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.001737] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.001909] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.001919] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.002095] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.002105] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.002179] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.002188] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.002300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.002310] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.002539] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.002549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.002727] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.002736] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.002849] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.002859] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.002966] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.002976] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.003208] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.003218] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.003331] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.003342] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.003568] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.003578] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.003684] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.003693] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.003886] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.003898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.004170] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.004180] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.004287] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.004297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.004420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.004430] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.004658] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.004668] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.004766] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.004775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.004947] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.004957] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.005054] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.005063] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.005182] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.005192] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.005368] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.005378] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.005570] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.005600] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.005818] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.005847] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.006049] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.440 [2024-07-15 18:12:48.006078] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.440 qpair failed and we were unable to recover it. 00:26:54.440 [2024-07-15 18:12:48.006215] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.006256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.006540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.006570] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.006759] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.006768] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.006962] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.006992] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.007202] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.007253] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.007478] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.007507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.007675] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.007704] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.007986] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.008015] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.008306] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.008336] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.008578] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.008608] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.008830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.008859] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.009081] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.009110] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.009320] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.009330] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.009565] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.009594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.009891] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.009921] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.010135] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.010164] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.010391] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.010421] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.010625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.010655] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.010927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.010957] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.011199] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.011237] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.011443] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.011472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.011628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.011638] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.011864] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.011874] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.012037] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.012058] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.012219] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.012232] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.012436] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.012465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.012675] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.012704] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.012975] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.013010] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.013270] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.013300] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.013519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.013529] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.013699] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.013709] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.013950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.013980] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.014131] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.014160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.014471] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.014481] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.014659] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.014669] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.014911] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.014941] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.015154] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.015183] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.015405] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.015435] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.015650] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.015680] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.015966] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.015994] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.016269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.016299] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.016446] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.016456] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.016568] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.016578] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.016802] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.016812] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.016984] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.016993] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.017248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.017279] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.017412] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.017422] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.017674] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.017684] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.017791] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.017801] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.018079] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.018089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.018300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.018310] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.018411] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.018420] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.018582] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.018592] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.441 qpair failed and we were unable to recover it. 00:26:54.441 [2024-07-15 18:12:48.018853] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.441 [2024-07-15 18:12:48.018883] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.019119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.019150] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.019413] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.019423] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.019597] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.019607] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.019793] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.019823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.020037] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.020067] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.020280] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.020310] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.020521] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.020549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.020760] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.020790] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.020937] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.020965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.021219] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.021259] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.021558] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.021568] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.021741] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.021751] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.021924] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.021934] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.022110] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.022132] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.022412] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.022443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.022603] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.022632] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.022758] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.022767] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.022946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.022956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.023075] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.023085] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.023248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.023257] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.023418] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.023461] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.023663] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.023692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.023981] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.024011] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.024156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.024185] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.024509] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.024539] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.024755] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.024776] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.025028] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.025037] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.025280] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.025290] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.025411] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.025421] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.025669] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.025679] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.025918] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.025947] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.026213] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.026269] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.026564] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.026574] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.026743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.026753] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.026929] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.026938] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.027213] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.027252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.027521] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.027550] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.027769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.027798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.028075] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.028105] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.028425] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.028456] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.028620] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.028649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.028848] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.028857] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.029027] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.029058] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.029269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.442 [2024-07-15 18:12:48.029298] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.442 qpair failed and we were unable to recover it. 00:26:54.442 [2024-07-15 18:12:48.029458] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.029487] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.029653] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.029683] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.029949] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.029979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.030181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.030210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.030420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.030450] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.030670] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.030699] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.030931] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.030940] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.031135] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.031144] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.031321] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.031332] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.031447] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.031458] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.031638] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.031648] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.031773] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.031782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.032094] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.032124] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.032346] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.032376] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.032590] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.032620] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.032912] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.032921] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.033084] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.033093] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.033266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.033297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.033608] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.033638] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.033789] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.033818] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.034036] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.034064] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.034250] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.034280] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.034504] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.034532] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.034685] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.034714] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.035008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.035037] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.035326] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.035357] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.035502] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.035512] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.035684] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.035694] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.035900] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.035928] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.036117] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.036145] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.036358] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.036387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.036549] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.036559] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.036801] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.036830] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.037052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.037081] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.037297] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.037327] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.037459] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.037468] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.037576] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.037586] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.037757] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.037767] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.037961] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.037989] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.038136] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.038165] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.038391] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.038422] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.038633] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.038643] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.038841] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.038851] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.039024] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.039034] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.039292] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.039323] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.039478] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.039507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.039657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.039685] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.039967] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.443 [2024-07-15 18:12:48.039977] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.443 qpair failed and we were unable to recover it. 00:26:54.443 [2024-07-15 18:12:48.040135] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.040145] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.040373] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.040387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.040562] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.040572] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.040741] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.040751] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.040939] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.040949] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.041109] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.041119] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.041289] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.041299] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.041500] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.041510] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.041603] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.041612] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.041804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.041814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.041981] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.041992] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.042160] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.042170] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.042241] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.042251] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.042350] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.042360] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.042472] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.042483] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.042712] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.042721] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.042999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.043029] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.043298] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.043333] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.043443] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.043453] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.043614] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.043624] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.043741] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.043750] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.043908] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.043919] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.044097] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.044107] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.044281] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.044291] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.044401] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.044412] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.044579] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.044589] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.044701] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.044710] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.044809] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.044819] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.044954] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.044988] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.045178] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.045194] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.045457] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.045472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.045589] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.045603] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.045770] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.045784] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.045956] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.045970] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.046168] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.046203] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.046454] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.046484] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.046710] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.046740] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.046905] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.046935] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.047082] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.047112] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.047315] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.047346] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.047622] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.047652] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.047798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.047837] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.047989] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.048019] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.048157] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.048187] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.048417] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.048449] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.048605] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.048618] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.048814] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.048843] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.444 [2024-07-15 18:12:48.049134] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.444 [2024-07-15 18:12:48.049164] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.444 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.049392] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.049406] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.049664] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.049678] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.049795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.049809] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.050061] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.050074] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.050325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.050357] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.050520] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.050549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.050767] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.050797] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.051005] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.051036] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.051242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.051273] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.051453] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.051483] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.051721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.051752] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.052009] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.052023] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.052223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.052241] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.052400] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.052413] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.052520] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.052534] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.052732] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.052745] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.052930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.052944] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.053095] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.053124] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.053355] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.053387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.053546] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.053576] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.053841] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.053909] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.054145] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.054179] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.054416] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.054449] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.054739] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.054769] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.054904] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.054914] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.055082] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.055091] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.055317] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.055327] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.055428] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.055437] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.055554] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.055564] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.055789] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.055799] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.055892] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.055901] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.056079] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.056089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.056194] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.056204] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.056314] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.056327] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.056560] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.056590] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.056817] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.056846] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.056994] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.057023] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.057190] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.057219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.057382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.057411] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.057618] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.057646] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.057876] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.057904] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.058174] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.058203] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.058517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.058548] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.058765] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.058795] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.058993] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.059003] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.059096] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.059105] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.059236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.059246] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.059472] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.059482] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.059742] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.059751] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.059974] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.059984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.060100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.060110] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.060215] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.060228] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.060476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.060486] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.060598] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.060608] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.445 [2024-07-15 18:12:48.060848] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.445 [2024-07-15 18:12:48.060858] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.445 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.061052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.061062] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.061227] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.061256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.061476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.061505] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.061724] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.061754] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.061920] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.061929] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.062094] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.062125] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.062352] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.062383] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.062645] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.062655] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.062838] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.062848] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.062963] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.062973] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.063089] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.063099] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.063266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.063277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.063471] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.063481] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.063640] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.063650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.063824] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.063834] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.063930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.063959] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.064188] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.064217] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.064537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.064567] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.064768] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.064803] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.064960] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.064989] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.065205] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.065242] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.065540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.065569] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.065830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.065840] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.065963] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.065973] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.066196] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.066206] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.066377] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.066387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.066549] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.066559] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.066784] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.066815] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.067046] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.067075] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.067270] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.067301] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.067466] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.067476] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.067638] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.067664] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.067892] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.067922] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.068068] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.068098] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.068316] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.068346] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.068579] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.068589] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.068831] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.068841] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.068997] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.069007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.069192] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.069202] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.069465] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.069494] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.069796] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.069825] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.070015] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.070025] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.070139] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.070149] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.070313] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.070323] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.070447] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.070457] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.070639] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.070649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.070830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.446 [2024-07-15 18:12:48.070859] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.446 qpair failed and we were unable to recover it. 00:26:54.446 [2024-07-15 18:12:48.071014] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.071043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.071271] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.071301] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.071449] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.071459] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.071540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.071549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.071726] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.071736] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.071900] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.071911] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.072069] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.072079] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.072165] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.072174] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.072423] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.072433] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.072543] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.072553] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.072671] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.072680] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.072781] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.072793] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.072949] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.072958] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.073132] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.073142] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.073237] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.073247] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.073364] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.073374] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.073484] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.073494] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.073654] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.073663] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.073906] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.073935] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.074138] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.074167] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.074303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.074334] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.074600] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.074610] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.074809] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.074819] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.074915] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.074924] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.075173] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.075183] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.075390] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.075401] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.075521] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.075531] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.075782] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.075812] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.075934] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.075964] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.076167] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.076196] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.076505] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.076535] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.076765] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.076795] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.077064] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.077093] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.077222] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.077262] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.077420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.077449] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.077711] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.077741] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.077887] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.077916] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.078184] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.078213] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.078445] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.078480] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.078633] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.078642] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.078879] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.078908] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.079138] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.079168] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.079323] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.079353] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.079555] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.079584] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.079851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.079880] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.080099] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.080128] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.080353] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.080384] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.080597] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.080625] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.447 qpair failed and we were unable to recover it. 00:26:54.447 [2024-07-15 18:12:48.080902] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.447 [2024-07-15 18:12:48.080911] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.081011] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.081021] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.081200] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.081210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.081358] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.081388] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.081686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.081716] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.081917] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.081926] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.082103] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.082113] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.082353] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.082363] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.082535] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.082545] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.082707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.082748] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.082985] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.083015] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.083238] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.083268] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.083504] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.083534] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.083667] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.083697] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.083898] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.083908] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.084134] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.084143] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.084328] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.084338] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.084581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.084610] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.084748] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.084778] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.084917] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.084946] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.085121] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.085150] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.085322] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.085352] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.085592] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.085622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.085890] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.085920] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.086140] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.086170] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.086393] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.086423] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.086707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.086716] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.086874] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.086885] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.087081] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.087110] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.087319] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.087349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.087541] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.087575] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.087763] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.087773] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.087930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.087940] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.088029] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.088038] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.088320] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.088350] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.088489] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.088518] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.088736] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.088765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.089033] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.089062] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.089283] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.089313] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.089534] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.089563] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.089838] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.089848] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.089952] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.089962] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.090056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.090065] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.090240] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.090250] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.090488] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.090519] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.090733] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.090763] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.091021] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.091050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.091255] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.448 [2024-07-15 18:12:48.091285] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.448 qpair failed and we were unable to recover it. 00:26:54.448 [2024-07-15 18:12:48.091421] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.091451] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.091635] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.091664] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.091885] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.091914] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.092133] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.092162] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.092375] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.092406] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.092542] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.092571] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.092728] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.092738] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.092848] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.092858] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.092976] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.092986] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.093233] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.093264] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.093472] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.093502] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.093713] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.093742] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.093958] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.093968] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.094066] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.094078] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.094239] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.094249] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.094405] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.094415] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.094571] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.094581] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.094705] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.094715] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.094888] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.094897] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.095067] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.095076] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.095319] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.095349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.095562] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.095592] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.095792] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.095825] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.096044] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.096054] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.096234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.096244] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.096419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.096428] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.096627] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.096656] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.096872] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.096901] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.097110] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.097140] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.097386] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.097416] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.097701] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.097730] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.097977] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.097987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.098089] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.098099] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.098202] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.098211] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.098458] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.098468] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.098654] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.098664] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.098840] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.098870] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.099159] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.099188] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.099411] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.099441] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.099611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.099641] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.099927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.099937] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.100180] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.449 [2024-07-15 18:12:48.100190] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.449 qpair failed and we were unable to recover it. 00:26:54.449 [2024-07-15 18:12:48.100429] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.100439] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.100553] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.100563] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.100737] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.100747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.100916] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.100926] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.101124] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.101134] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.101385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.101420] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.101522] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.101532] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.101653] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.101662] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.101890] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.101919] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.102057] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.102087] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.102354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.102383] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.102538] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.102548] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.102716] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.102752] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.102994] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.103023] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.103183] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.103212] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.103376] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.103406] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.103625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.103654] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.103891] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.103920] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.104064] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.104093] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.104313] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.104343] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.104562] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.104596] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.104805] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.104833] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.105093] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.105103] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.105378] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.105388] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.105559] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.105569] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.105662] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.105671] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.105830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.105840] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.106013] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.106023] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.106203] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.106252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.106476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.106506] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.106642] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.106671] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.106902] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.106912] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.107048] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.107058] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.107239] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.107250] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.107332] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.107341] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.107448] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.107457] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.107705] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.107715] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.107824] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.107834] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.107946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.107956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.108055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.108064] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.108136] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.108145] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.108371] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.108382] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.108537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.108547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.108728] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.108738] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.108984] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.109014] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.109217] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.109264] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.109528] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.109558] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.109875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.109885] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.110136] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.110146] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.110323] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.110333] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.110517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.110527] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.110728] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.110737] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.110900] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.110910] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.111037] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.111047] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.111162] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.111171] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.111347] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.450 [2024-07-15 18:12:48.111357] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.450 qpair failed and we were unable to recover it. 00:26:54.450 [2024-07-15 18:12:48.111448] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.111457] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.111577] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.111587] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.111723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.111733] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.111915] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.111945] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.112159] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.112193] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.112444] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.112475] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.112599] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.112628] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.112799] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.112828] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.113009] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.113038] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.113308] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.113338] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.113559] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.113588] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.113759] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.113788] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.114073] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.114102] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.114252] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.114282] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.114504] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.114533] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.114742] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.114762] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.114940] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.114950] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.115132] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.115162] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.115423] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.115453] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.115724] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.115754] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.116045] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.116075] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.116293] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.116341] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.116652] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.116682] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.116896] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.116926] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.117119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.117128] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.117303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.117313] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.117426] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.117436] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.117627] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.117637] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.117744] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.117754] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.117877] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.117888] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.118122] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.118152] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.118378] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.118408] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.118616] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.118626] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.118792] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.118812] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.118909] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.118918] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.119132] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.119142] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.119319] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.119329] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.119491] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.119501] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.119595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.119604] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.119709] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.119719] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.119836] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.119846] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.119968] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.119978] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.120146] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.120156] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.120355] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.120365] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.120560] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.120572] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.120670] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.120679] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.120871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.120901] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.121178] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.121207] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.121425] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.121455] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.121718] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.121728] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.121927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.121936] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.122122] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.122131] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.122320] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.122330] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.122507] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.122537] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.122779] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.122808] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.122965] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.122994] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.451 qpair failed and we were unable to recover it. 00:26:54.451 [2024-07-15 18:12:48.123267] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.451 [2024-07-15 18:12:48.123297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.452 qpair failed and we were unable to recover it. 00:26:54.452 [2024-07-15 18:12:48.123514] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.452 [2024-07-15 18:12:48.123543] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.452 qpair failed and we were unable to recover it. 00:26:54.452 [2024-07-15 18:12:48.123713] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.452 [2024-07-15 18:12:48.123743] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.452 qpair failed and we were unable to recover it. 00:26:54.452 [2024-07-15 18:12:48.123908] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.452 [2024-07-15 18:12:48.123936] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.452 qpair failed and we were unable to recover it. 00:26:54.452 [2024-07-15 18:12:48.124071] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.452 [2024-07-15 18:12:48.124081] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.452 qpair failed and we were unable to recover it. 00:26:54.452 [2024-07-15 18:12:48.124189] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.452 [2024-07-15 18:12:48.124199] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.452 qpair failed and we were unable to recover it. 00:26:54.452 [2024-07-15 18:12:48.124380] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.452 [2024-07-15 18:12:48.124390] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.452 qpair failed and we were unable to recover it. 00:26:54.452 [2024-07-15 18:12:48.124541] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.452 [2024-07-15 18:12:48.124551] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.452 qpair failed and we were unable to recover it. 00:26:54.452 [2024-07-15 18:12:48.124722] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.452 [2024-07-15 18:12:48.124732] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.452 qpair failed and we were unable to recover it. 00:26:54.452 [2024-07-15 18:12:48.124853] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.452 [2024-07-15 18:12:48.124863] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.452 qpair failed and we were unable to recover it. 00:26:54.452 [2024-07-15 18:12:48.125022] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.452 [2024-07-15 18:12:48.125032] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.452 qpair failed and we were unable to recover it. 00:26:54.452 [2024-07-15 18:12:48.125131] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.452 [2024-07-15 18:12:48.125140] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.452 qpair failed and we were unable to recover it. 00:26:54.452 [2024-07-15 18:12:48.125320] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.452 [2024-07-15 18:12:48.125330] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.452 qpair failed and we were unable to recover it. 00:26:54.452 [2024-07-15 18:12:48.125512] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.452 [2024-07-15 18:12:48.125541] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.452 qpair failed and we were unable to recover it. 00:26:54.452 [2024-07-15 18:12:48.125752] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.452 [2024-07-15 18:12:48.125782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.452 qpair failed and we were unable to recover it. 00:26:54.452 [2024-07-15 18:12:48.126018] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.452 [2024-07-15 18:12:48.126048] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.452 qpair failed and we were unable to recover it. 00:26:54.452 [2024-07-15 18:12:48.126262] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.452 [2024-07-15 18:12:48.126293] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.452 qpair failed and we were unable to recover it. 00:26:54.452 [2024-07-15 18:12:48.126615] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.452 [2024-07-15 18:12:48.126655] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.452 qpair failed and we were unable to recover it. 00:26:54.742 [2024-07-15 18:12:48.126820] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.742 [2024-07-15 18:12:48.126832] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.742 qpair failed and we were unable to recover it. 00:26:54.742 [2024-07-15 18:12:48.127037] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.742 [2024-07-15 18:12:48.127067] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.742 qpair failed and we were unable to recover it. 00:26:54.742 [2024-07-15 18:12:48.127187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.742 [2024-07-15 18:12:48.127217] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.742 qpair failed and we were unable to recover it. 00:26:54.742 [2024-07-15 18:12:48.127372] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.742 [2024-07-15 18:12:48.127402] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.742 qpair failed and we were unable to recover it. 00:26:54.742 [2024-07-15 18:12:48.127639] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.742 [2024-07-15 18:12:48.127669] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.742 qpair failed and we were unable to recover it. 00:26:54.742 [2024-07-15 18:12:48.127811] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.742 [2024-07-15 18:12:48.127822] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.742 qpair failed and we were unable to recover it. 00:26:54.742 [2024-07-15 18:12:48.128068] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.742 [2024-07-15 18:12:48.128098] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.742 qpair failed and we were unable to recover it. 00:26:54.742 [2024-07-15 18:12:48.128259] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.742 [2024-07-15 18:12:48.128294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.742 qpair failed and we were unable to recover it. 00:26:54.742 [2024-07-15 18:12:48.128511] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.742 [2024-07-15 18:12:48.128543] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.742 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.128755] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.128765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.128878] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.128890] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.129061] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.129070] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.129341] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.129351] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.129560] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.129570] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.129809] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.129819] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.129981] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.129991] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.130191] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.130201] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.130363] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.130373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.130480] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.130489] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.130595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.130606] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.130724] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.130734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.130835] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.130844] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.131070] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.131081] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.131181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.131190] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.131302] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.131312] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.131410] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.131418] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.131519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.131528] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.131750] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.131760] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.131872] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.131882] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.131994] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.132004] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.132166] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.132177] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.132290] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.132301] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.132458] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.132468] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.132643] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.132653] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.132855] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.132884] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.133098] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.133127] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.133284] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.133315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.133532] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.133563] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.133795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.133805] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.133908] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.133917] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.134022] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.134031] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.134208] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.134218] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.134392] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.134402] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.134579] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.134588] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.134695] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.134706] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.134933] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.134943] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.135051] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.135061] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.135174] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.135183] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.135341] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.135351] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.135438] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.135447] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.135619] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.135631] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.135857] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.135867] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.135977] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.135987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.136132] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.136142] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.136301] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.136311] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.136494] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.136504] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.136673] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.136682] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.136784] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.136801] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.136898] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.136910] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.137074] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.137084] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.137174] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.137184] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.137298] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.137308] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.137483] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.137493] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.137604] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.137614] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.137711] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.137720] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.137883] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.137894] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.138121] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.138131] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.138322] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.138333] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.138435] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.138445] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.138619] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.138629] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.138711] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.138720] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.138932] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.138942] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.139102] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.139112] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.139208] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.139217] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.139337] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.743 [2024-07-15 18:12:48.139347] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.743 qpair failed and we were unable to recover it. 00:26:54.743 [2024-07-15 18:12:48.139522] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.139532] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.139690] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.139700] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.139878] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.139889] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.140056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.140066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.140161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.140170] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.140436] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.140446] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.140605] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.140614] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.140792] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.140802] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.140965] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.140974] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.141134] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.141144] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.141242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.141252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.141487] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.141517] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.141785] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.141814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.142051] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.142060] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.142227] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.142238] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.142366] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.142378] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.142585] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.142595] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.142707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.142718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.142812] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.142821] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.143003] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.143013] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.143256] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.143267] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.143388] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.143398] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.143466] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.143475] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.143589] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.143597] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.143809] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.143819] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.143979] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.143988] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.144145] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.144155] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.144414] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.144424] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.144534] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.144544] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.144743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.144753] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.144937] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.144947] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.145037] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.145046] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.145269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.145280] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.145379] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.145389] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.145514] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.145524] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.145615] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.145625] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.145860] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.145870] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.146033] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.146043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.146213] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.146223] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.146406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.146416] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.146520] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.146529] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.146628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.146638] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.146822] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.146831] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.147027] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.147036] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.147147] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.147157] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.147327] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.147337] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.147439] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.147449] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.147623] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.147633] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.147822] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.147832] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.147948] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.147958] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.148083] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.148093] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.148193] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.148203] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.148367] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.148377] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.148471] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.148481] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.148546] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.148555] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.148723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.148734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.148895] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.148905] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.149016] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.149026] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.149205] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.149215] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.149392] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.149402] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.149517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.149527] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.149689] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.149699] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.149934] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.149963] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.150237] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.150267] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.744 [2024-07-15 18:12:48.150427] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.744 [2024-07-15 18:12:48.150457] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.744 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.150614] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.150643] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.150870] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.150899] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.151137] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.151167] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.151310] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.151341] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.151588] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.151618] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.151817] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.151846] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.152063] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.152072] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.152259] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.152270] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.152379] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.152389] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.152559] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.152568] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.152687] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.152697] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.152855] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.152864] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.153024] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.153034] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.153140] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.153150] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.153254] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.153263] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.153423] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.153432] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.153625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.153635] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.153814] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.153825] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.153947] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.153957] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.154064] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.154074] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.154178] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.154188] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.154289] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.154298] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.154459] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.154469] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.154623] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.154633] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.154793] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.154803] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.154927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.154937] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.155099] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.155109] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.155299] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.155309] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.155468] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.155478] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.155602] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.155612] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.155721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.155733] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.155958] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.155968] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.156068] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.156078] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.156249] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.156264] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.156369] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.156378] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.156505] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.156515] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.156721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.156731] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.156888] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.156898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.157071] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.157080] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.157264] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.157274] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.157460] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.157469] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.157626] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.157636] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.157747] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.157757] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.157865] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.157874] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.158058] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.158068] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.158259] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.158289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.158426] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.158456] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.158678] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.158707] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.158926] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.158956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.159147] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.159156] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.159260] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.159270] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.159546] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.159555] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.159832] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.159860] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.159997] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.160027] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.160195] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.160258] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.160483] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.745 [2024-07-15 18:12:48.160512] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.745 qpair failed and we were unable to recover it. 00:26:54.745 [2024-07-15 18:12:48.160661] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.160690] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.160934] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.160968] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.161209] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.161231] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.161341] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.161355] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.161544] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.161558] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.161738] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.161752] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.161936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.161949] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.162078] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.162092] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.162211] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.162230] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.162331] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.162345] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.162581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.162595] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.162763] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.162776] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.162946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.162959] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.163078] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.163092] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.163200] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.163218] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.163325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.163339] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.163529] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.163542] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.163777] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.163791] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.163978] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.163992] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.164162] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.164175] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.164434] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.164448] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.164566] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.164579] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.164689] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.164703] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.164873] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.164887] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.165134] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.165148] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.165332] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.165347] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.165486] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.165520] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.165800] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.165829] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.166040] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.166071] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.166306] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.166319] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.166504] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.166518] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.166797] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.166810] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.166985] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.166999] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.167180] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.167210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.167449] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.167479] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.167693] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.167731] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.167982] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.167995] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.168168] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.168182] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.168422] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.168436] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.168668] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.168698] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.168899] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.168929] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.169169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.169181] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.169363] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.169374] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.169544] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.169554] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.169728] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.169738] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.169834] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.169844] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.170014] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.170023] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.170183] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.170193] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.170371] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.170382] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.170562] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.170572] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.170748] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.170758] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.170925] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.170935] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.171099] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.171109] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.171280] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.171290] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.171448] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.171460] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.171554] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.171563] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.171715] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.171724] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.171896] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.171906] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.171996] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.172006] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.172184] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.172194] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.172437] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.172447] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.172552] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.172561] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.172656] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.172664] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.172864] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.172874] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.173020] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.173030] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.173140] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.746 [2024-07-15 18:12:48.173149] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.746 qpair failed and we were unable to recover it. 00:26:54.746 [2024-07-15 18:12:48.173382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.173391] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.173484] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.173494] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.173677] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.173687] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.173846] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.173856] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.174098] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.174107] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.174229] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.174239] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.174539] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.174569] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.174707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.174736] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.174951] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.174981] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.175172] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.175183] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.175371] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.175382] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.175560] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.175570] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.175744] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.175754] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.175927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.175937] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.176079] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.176089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.176331] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.176365] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.176578] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.176608] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.176812] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.176841] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.177002] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.177015] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.177121] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.177135] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.177252] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.177267] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.177505] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.177518] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.177693] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.177706] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.177937] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.177951] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.178132] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.178146] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.178263] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.178277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.178447] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.178461] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.178628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.178642] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.178809] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.178823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.178927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.178941] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.179043] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.179056] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.179293] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.179307] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.179510] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.179524] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.179689] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.179702] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.179806] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.179820] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.180036] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.180050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.180257] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.180271] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.180441] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.180455] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.180555] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.180569] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.180656] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.180668] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.180842] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.180856] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.181052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.181066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.181171] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.181184] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.181296] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.181311] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.181514] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.181527] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.181685] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.181698] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.181899] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.181913] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.182085] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.182098] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.182287] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.182300] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.182490] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.182503] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.182680] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.182694] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.182861] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.182874] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.182986] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.182999] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.183173] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.183186] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.183315] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.183329] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.183505] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.183521] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.183688] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.183701] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.183872] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.183886] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.183997] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.184011] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.184182] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.184195] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.184362] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.184376] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.184492] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.184505] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.184749] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.184763] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.184951] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.184965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.185205] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.185219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.185429] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.747 [2024-07-15 18:12:48.185459] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.747 qpair failed and we were unable to recover it. 00:26:54.747 [2024-07-15 18:12:48.185663] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.185692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.185901] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.185930] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.186076] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.186090] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.186355] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.186369] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.186620] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.186634] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.186808] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.186823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.186939] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.186953] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.187083] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.187096] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.187245] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.187259] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.187446] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.187475] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.187760] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.187789] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.188006] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.188044] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.188161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.188174] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.188428] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.188443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.188573] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.188586] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.188848] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.188861] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.188965] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.188978] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.189153] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.189167] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.189354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.189368] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.189628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.189657] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.189957] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.189986] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.190200] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.190214] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.190342] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.190356] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.190569] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.190582] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.190773] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.190787] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.190961] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.190974] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.191105] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.191118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.191292] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.191306] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.191489] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.191503] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.191635] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.191651] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.191765] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.191778] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.191960] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.191974] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.192087] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.192100] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.192281] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.192295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.192451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.192464] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.192704] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.192718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.192974] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.192987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.193088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.193101] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.193216] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.193233] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.193403] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.193416] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.193619] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.193633] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.193808] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.193821] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.193933] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.193947] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.194076] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.194090] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.194196] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.194210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.194394] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.194408] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.194679] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.194693] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.194858] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.194871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.195039] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.195053] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.195233] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.195247] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.195435] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.195448] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.195657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.195671] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.195850] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.195864] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.196040] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.196053] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.196244] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.196259] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.196434] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.196448] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.196712] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.196725] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.196964] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.196977] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.748 [2024-07-15 18:12:48.197146] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.748 [2024-07-15 18:12:48.197159] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.748 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.197341] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.197356] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.197590] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.197604] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.197788] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.197801] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.197924] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.197937] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.198052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.198066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.198239] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.198254] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.198451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.198481] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.198630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.198659] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.198812] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.198842] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.198999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.199028] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.199228] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.199244] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.199419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.199432] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.199623] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.199636] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.199875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.199888] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.200145] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.200159] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.200370] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.200384] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.200561] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.200574] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.200743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.200756] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.200964] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.200978] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.201090] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.201104] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.201299] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.201313] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.201442] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.201455] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.201609] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.201622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.201835] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.201849] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.202080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.202094] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.202261] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.202275] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.202501] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.202530] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.202798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.202839] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.202943] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.202956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.203064] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.203077] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.203284] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.203297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.203435] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.203449] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.203566] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.203580] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.203681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.203694] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.203865] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.203879] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.204004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.204017] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.204106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.204119] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.204300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.204314] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.204451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.204465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.204563] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.204575] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.204791] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.204805] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.204915] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.204929] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.205035] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.205049] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.205260] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.205275] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.205440] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.205454] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.205583] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.205597] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.205776] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.205790] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.205969] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.205982] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.206085] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.206099] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.206235] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.206249] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.206368] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.206383] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.206488] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.206502] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.206714] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.206727] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.206833] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.206847] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.207028] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.207042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.207237] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.207251] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.207442] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.207456] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.207585] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.207598] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.207826] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.207840] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.208038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.208052] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.208287] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.208301] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.208500] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.208514] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.208683] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.208697] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.208875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.749 [2024-07-15 18:12:48.208888] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.749 qpair failed and we were unable to recover it. 00:26:54.749 [2024-07-15 18:12:48.209020] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.209034] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.209151] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.209164] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.209275] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.209289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.209403] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.209417] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.209686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.209700] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.209802] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.209816] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.209986] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.210000] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.210188] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.210201] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.210411] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.210424] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.210552] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.210566] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.210734] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.210747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.210850] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.210864] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.211037] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.211050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.211275] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.211289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.211472] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.211486] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.211603] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.211617] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.211787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.211800] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.211969] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.211982] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.212262] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.212292] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.212418] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.212448] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.212743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.212773] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.213035] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.213048] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.213217] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.213236] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.213440] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.213453] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.213659] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.213673] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.213908] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.213922] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.214051] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.214067] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.214187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.214200] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.214382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.214397] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.214505] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.214519] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.214692] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.214705] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.214884] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.214897] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.215006] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.215019] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.215189] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.215202] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.215373] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.215387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.215579] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.215592] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.215707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.215721] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.215952] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.215966] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.216088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.216101] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.216238] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.216252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.216445] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.216459] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.216629] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.216643] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.216947] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.216961] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.217143] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.217156] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.217270] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.217283] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.217392] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.217406] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.217607] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.217621] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.217741] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.217754] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.217954] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.217968] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.218152] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.218166] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.218345] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.218359] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.218477] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.218490] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.218603] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.218616] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.218745] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.218759] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.218945] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.218958] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.219078] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.219091] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.219272] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.219286] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.219464] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.219494] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.219706] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.219736] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.750 qpair failed and we were unable to recover it. 00:26:54.750 [2024-07-15 18:12:48.219935] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.750 [2024-07-15 18:12:48.219965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.220236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.220250] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.220453] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.220466] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.220570] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.220583] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.220790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.220803] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.221001] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.221015] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.221204] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.221218] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.221410] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.221427] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.221526] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.221538] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.221713] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.221727] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.221829] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.221842] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.222102] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.222116] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.222363] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.222377] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.222555] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.222568] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.222746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.222777] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.223006] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.223035] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.223248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.223278] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.223567] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.223581] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.223770] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.223783] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.223969] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.223982] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.224102] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.224116] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.224247] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.224261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.224428] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.224442] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.224547] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.224561] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.224752] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.224766] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.224945] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.224959] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.225138] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.225152] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.225384] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.225398] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.225567] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.225581] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.225770] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.225784] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.226075] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.226089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.226215] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.226234] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.226408] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.226422] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.226553] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.226566] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.226734] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.226747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.226898] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.226927] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.227214] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.227256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.227526] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.227555] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.227841] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.227854] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.228067] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.228081] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.228349] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.228363] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.228540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.228570] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.228787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.228816] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.229029] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.229058] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.229288] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.229303] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.229474] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.229488] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.229657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.229670] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.229838] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.229854] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.230055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.230068] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.230239] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.230253] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.230440] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.230470] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.230738] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.230768] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.231013] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.231055] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.231314] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.231328] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.231437] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.231450] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.231629] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.231642] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.231822] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.231836] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.231945] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.231959] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.232196] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.232210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.232335] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.232355] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.232457] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.232469] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.232678] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.232691] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.232808] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.232822] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.232995] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.233008] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.751 [2024-07-15 18:12:48.233111] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.751 [2024-07-15 18:12:48.233124] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.751 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.233291] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.233305] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.233497] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.233511] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.233720] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.233733] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.234082] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.234111] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.234310] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.234340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.234635] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.234664] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.234862] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.234891] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.235043] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.235072] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.235287] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.235322] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.235503] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.235517] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.235809] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.235838] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.236039] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.236068] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.236336] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.236366] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.236517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.236546] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.236769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.236799] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.236992] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.237005] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.237248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.237279] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.237572] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.237602] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.237782] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.237810] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.238011] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.238040] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.238258] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.238288] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.238561] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.238591] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.238813] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.238848] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.239004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.239033] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.239187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.239216] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.239432] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.239463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.239757] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.239787] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.240055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.240084] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.240305] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.240319] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.240553] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.240566] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.240699] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.240712] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.240971] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.240984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.241097] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.241110] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.241229] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.241243] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.241432] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.241446] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.241573] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.241586] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.241718] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.241733] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.241909] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.241922] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.242149] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.242162] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.242281] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.242295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.242406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.242420] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.242538] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.242552] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.242672] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.242686] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.242855] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.242868] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.242970] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.242983] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.243099] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.243112] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.243247] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.243261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.243496] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.243510] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.243683] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.243696] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.243803] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.243817] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.243919] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.243933] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.244101] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.244115] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.244373] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.244387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.244582] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.244595] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.244802] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.244831] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.245038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.245068] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.245216] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.245254] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.245545] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.245558] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.245790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.245804] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.245963] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.245977] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.246090] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.246103] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.246341] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.246356] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.246526] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.246542] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.246655] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.246669] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.246902] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.246915] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.247040] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.247053] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.247234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.247249] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.247368] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.247382] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.752 [2024-07-15 18:12:48.247563] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.752 [2024-07-15 18:12:48.247576] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.752 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.247688] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.247702] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.247826] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.247840] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.248039] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.248052] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.248142] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.248154] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.248407] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.248421] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.248521] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.248538] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.248657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.248670] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.248844] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.248858] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.248988] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.249001] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.249114] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.249128] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.249235] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.249248] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.249415] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.249428] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.249593] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.249606] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.249779] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.249792] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.249971] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.249984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.250195] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.250208] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.250411] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.250425] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.250600] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.250629] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.250783] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.250812] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.251018] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.251047] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.251205] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.251245] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.251519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.251549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.251791] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.251821] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.252020] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.252033] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.252146] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.252160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.252393] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.252408] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.252531] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.252545] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.252657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.252671] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.252852] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.252866] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.253067] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.253080] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.253222] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.253240] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.253407] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.253421] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.253611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.253624] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.253795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.253810] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.254041] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.254055] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.254318] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.254348] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.254584] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.254614] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.254747] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.254776] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.254950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.254979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.255221] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.255237] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.255409] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.255422] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.255540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.255554] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.255790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.255804] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.255930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.255943] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.256112] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.256125] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.256272] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.256287] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.256473] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.256515] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.256724] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.256753] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.257035] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.257064] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.257280] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.257310] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.257551] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.257565] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.257799] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.257812] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.258070] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.258084] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.258265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.258279] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.258543] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.258556] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.258711] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.258725] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.258916] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.258946] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.259081] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.259111] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.259326] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.259356] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.259505] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.259536] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.259717] nvme_tcp.c: 327:nvme_tcp_qpair_set_recv_state: *ERROR*: The recv state of tqpair=0x1940000 is same with the state(5) to be set 00:26:54.753 [2024-07-15 18:12:48.260017] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.260085] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.753 qpair failed and we were unable to recover it. 00:26:54.753 [2024-07-15 18:12:48.260325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.753 [2024-07-15 18:12:48.260336] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.260473] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.260483] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.260666] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.260676] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.260787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.260797] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.260889] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.260899] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.261156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.261185] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.261504] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.261535] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.261672] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.261701] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.261842] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.261882] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.262057] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.262067] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.262321] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.262351] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.262581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.262612] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.262825] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.262855] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.263003] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.263032] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.263297] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.263327] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.263531] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.263560] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.263765] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.263794] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.264018] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.264048] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.264262] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.264293] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.264450] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.264480] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.264704] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.264734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.265028] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.265057] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.265211] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.265248] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.265478] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.265507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.265734] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.265763] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.265931] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.265966] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.266177] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.266187] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.266344] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.266355] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.266592] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.266621] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.266915] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.266945] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.267186] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.267196] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.267420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.267430] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.267537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.267547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.267721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.267730] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.267908] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.267917] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.268081] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.268102] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.268332] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.268363] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.268571] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.268600] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.268821] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.268850] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.269114] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.269124] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.269320] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.269331] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.269504] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.269533] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.269693] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.269722] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.269871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.269901] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.270207] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.270248] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.270406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.270436] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.270667] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.270697] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.270933] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.270965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.271099] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.271108] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.271331] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.271341] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.271568] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.271578] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.271740] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.271750] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.271941] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.271970] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.272183] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.272213] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.272429] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.272460] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.272683] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.272712] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.272941] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.272970] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.273237] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.273247] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.273413] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.273422] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.273601] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.273630] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.273919] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.273947] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.274167] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.274197] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.274415] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.274445] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.274712] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.274741] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.275057] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.275086] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.275326] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.275362] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.275542] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.275551] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.275739] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.275768] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.754 [2024-07-15 18:12:48.276008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.754 [2024-07-15 18:12:48.276038] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.754 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.276306] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.276337] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.276555] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.276566] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.276738] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.276748] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.276851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.276863] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.277115] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.277125] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.277303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.277313] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.277425] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.277434] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.277581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.277591] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.277763] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.277773] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.277886] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.277896] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.278091] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.278101] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.278230] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.278240] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.278399] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.278410] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.278652] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.278681] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.278838] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.278867] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.279020] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.279050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.279263] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.279273] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.279442] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.279452] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.279643] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.279672] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.279922] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.279951] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.280182] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.280211] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.280498] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.280528] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.280677] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.280707] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.280934] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.280964] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.281193] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.281223] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.281494] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.281504] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.281667] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.281677] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.281842] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.281852] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.282113] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.282142] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.282356] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.282386] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.282588] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.282617] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.282749] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.282779] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.282976] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.283006] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.283214] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.283254] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.283473] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.283502] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.283768] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.283798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.284087] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.284122] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.284348] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.284378] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.284569] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.284579] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.284850] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.284860] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.285030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.285040] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.285237] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.285267] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.285418] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.285448] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.285685] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.285714] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.286005] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.286034] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.286194] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.286223] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.286445] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.286455] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.286569] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.286578] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.286690] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.286700] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.286811] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.286821] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.287004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.287014] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.287133] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.287143] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.287253] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.287263] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.287422] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.287432] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.287639] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.287649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.287829] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.287838] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.288011] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.288021] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.288116] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.288126] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.288232] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.288243] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.288418] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.288429] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.288543] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.288553] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.288783] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.288793] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.288914] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.288924] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.289122] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.289156] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.289299] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.289315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.289492] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.289505] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.289625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.289638] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.289824] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.289837] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.290022] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.755 [2024-07-15 18:12:48.290036] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.755 qpair failed and we were unable to recover it. 00:26:54.755 [2024-07-15 18:12:48.290202] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.290216] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.290408] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.290422] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.290601] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.290614] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.290875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.290889] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.291075] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.291104] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.291405] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.291436] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.291748] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.291777] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.291886] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.291915] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.292135] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.292165] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.292381] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.292395] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.292576] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.292590] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.292851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.292880] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.293106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.293135] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.293397] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.293411] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.293659] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.293688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.293858] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.293887] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.294088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.294117] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.294369] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.294383] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.294574] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.294589] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.294760] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.294789] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.295009] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.295039] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.295363] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.295396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.295604] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.295633] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.295854] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.295883] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.296095] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.296124] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.296419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.296449] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.296747] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.296757] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.297005] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.297015] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.297232] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.297278] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.297460] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.297476] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.297689] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.297704] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.297890] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.297904] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.298109] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.298139] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.298378] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.298409] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.298620] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.298658] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.298894] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.298924] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.299149] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.299179] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.299348] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.299362] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.299595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.299609] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.299736] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.299749] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.299936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.299949] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.300161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.300175] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.300433] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.300449] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.300752] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.300768] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.300897] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.300910] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.301106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.301135] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.301276] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.301306] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.301507] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.301536] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.301784] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.301814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.302026] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.302055] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.302274] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.302306] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.302577] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.302606] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.302808] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.302838] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.303130] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.303160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.303440] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.303454] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.303653] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.303667] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.303876] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.303889] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.304091] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.304121] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.756 qpair failed and we were unable to recover it. 00:26:54.756 [2024-07-15 18:12:48.304422] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.756 [2024-07-15 18:12:48.304454] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.304677] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.304707] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.304871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.304901] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.305110] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.305126] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.305323] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.305337] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.305508] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.305537] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.305705] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.305735] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.306028] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.306058] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.306367] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.306381] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.306499] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.306513] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.306727] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.306741] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.307014] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.307028] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.307212] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.307229] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.307506] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.307535] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.307803] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.307833] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.307975] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.308004] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.308281] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.308295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.308533] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.308546] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.308729] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.308742] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.308948] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.308977] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.309255] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.309286] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.309423] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.309453] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.309636] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.309649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.309883] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.309897] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.310155] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.310169] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.310349] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.310364] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.310637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.310666] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.310835] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.310865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.310976] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.311005] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.311170] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.311184] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.311463] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.311494] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.311660] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.311690] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.311959] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.311988] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.312280] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.312310] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.312501] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.312531] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.312825] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.312854] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.313054] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.313084] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.313399] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.313429] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.313724] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.313753] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.314041] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.314071] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.314379] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.314393] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.314577] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.314590] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.314848] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.314862] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.314994] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.315010] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.315297] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.315328] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.315556] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.315585] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.315767] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.315796] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.315949] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.315978] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.316264] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.316295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.316569] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.316598] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.316800] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.316830] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.317045] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.317075] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.317342] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.317373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.317589] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.317619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.317831] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.317862] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.318066] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.318095] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.318270] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.318302] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.318613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.318627] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.318880] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.318893] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.319061] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.319075] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.319262] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.319292] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.319519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.319549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.319758] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.319788] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.320009] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.320039] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.320258] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.320289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.320574] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.320587] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.320701] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.320715] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.320895] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.320909] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.321086] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.757 [2024-07-15 18:12:48.321099] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.757 qpair failed and we were unable to recover it. 00:26:54.757 [2024-07-15 18:12:48.321305] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.321319] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.321501] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.321531] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.321679] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.321709] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.321851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.321881] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.322171] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.322201] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.322444] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.322475] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.322673] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.322703] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.322940] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.322969] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.323116] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.323130] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.323234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.323247] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.323442] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.323455] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.323657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.323671] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.323902] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.323915] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.324170] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.324183] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.324366] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.324383] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.324585] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.324615] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.324884] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.324914] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.325135] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.325148] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.325313] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.325326] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.325494] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.325507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.325677] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.325690] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.325785] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.325797] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.326064] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.326076] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.326207] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.326219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.326318] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.326331] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.326583] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.326595] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.326769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.326781] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.326950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.326963] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.327218] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.327235] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.327467] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.327480] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.327771] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.327783] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.327970] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.327984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.328241] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.328254] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.328444] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.328457] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.328574] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.328586] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.328703] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.328716] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.328916] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.328929] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.329162] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.329175] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.329318] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.329331] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.329532] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.329545] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.329644] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.329657] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.329848] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.329861] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.329993] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.330006] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.330181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.330194] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.330396] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.330409] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.330594] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.330606] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.330793] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.330806] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.330981] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.330994] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.331119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.331132] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.331313] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.331326] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.331545] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.331559] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.331690] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.331703] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.331875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.331889] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.332173] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.332202] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.332426] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.332462] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.332666] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.332680] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.332795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.332808] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.332980] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.332993] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.333170] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.333184] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.333309] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.333324] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.333491] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.333505] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.333624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.758 [2024-07-15 18:12:48.333637] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.758 qpair failed and we were unable to recover it. 00:26:54.758 [2024-07-15 18:12:48.333892] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.333905] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.334024] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.334037] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.334272] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.334286] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.334406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.334420] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.334547] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.334560] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.334800] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.334813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.334988] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.335002] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.335182] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.335195] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.335382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.335396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.335515] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.335529] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.335653] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.335667] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.335778] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.335792] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.335966] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.335980] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.336148] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.336161] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.336326] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.336340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.336427] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.336439] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.336558] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.336585] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.336789] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.336800] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.336965] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.336975] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.337140] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.337150] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.337288] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.337299] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.337501] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.337511] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.337638] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.337648] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.337831] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.337841] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.338018] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.338028] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.338135] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.338144] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.338255] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.338266] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.338444] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.338454] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.338568] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.338578] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.338675] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.338685] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.338855] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.338865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.338959] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.338969] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.339062] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.339074] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.339208] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.339218] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.339334] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.339344] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.339516] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.339525] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.339701] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.339730] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.339950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.339980] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.340100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.340130] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.340397] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.340407] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.340595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.340605] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.340700] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.340710] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.340877] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.340887] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.340982] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.340992] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.341152] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.341162] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.341265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.341275] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.341379] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.341388] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.341517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.341527] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.341698] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.341708] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.341877] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.341887] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.342048] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.342058] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.342218] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.342231] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.342460] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.342470] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.342639] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.342649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.342756] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.342765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.342945] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.342955] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.343129] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.343139] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.343247] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.343257] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.343434] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.343444] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.343615] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.343625] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.343785] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.343795] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.343889] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.343898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.344011] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.344020] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.344178] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.344188] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.344302] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.344312] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.344417] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.344427] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.344528] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.344538] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.344712] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.344722] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.344825] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.344835] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.345071] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.759 [2024-07-15 18:12:48.345081] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.759 qpair failed and we were unable to recover it. 00:26:54.759 [2024-07-15 18:12:48.345151] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.345160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.345340] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.345350] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.345519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.345530] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.345628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.345638] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.345871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.345880] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.346054] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.346064] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.346184] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.346194] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.346305] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.346315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.346551] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.346581] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.346875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.346904] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.347079] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.347108] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.347321] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.347351] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.347535] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.347544] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.347746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.347756] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.347933] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.347943] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.348042] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.348052] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.348230] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.348240] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.348471] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.348481] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.348653] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.348663] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.348901] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.348930] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.349086] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.349116] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.349428] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.349460] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.349644] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.349654] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.349804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.349814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.350053] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.350082] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.350362] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.350393] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.350620] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.350650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.350797] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.350827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.351118] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.351147] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.351326] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.351357] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.351581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.351611] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.351824] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.351853] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.352142] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.352172] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.352308] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.352340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.352468] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.352506] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.352669] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.352679] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.352853] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.352863] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.353025] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.353035] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.353204] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.353213] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.353455] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.353486] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.353707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.353737] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.353924] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.353953] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.354159] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.354194] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.354413] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.354445] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.354653] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.354682] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.354949] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.354979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.355191] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.355220] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.355385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.355396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.355519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.355529] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.355775] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.355785] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.355889] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.355899] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.356147] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.356157] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.356385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.356396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.356624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.356633] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.356803] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.356813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.357076] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.357106] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.357330] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.357340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.357457] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.357467] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.357574] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.357584] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.357769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.357779] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.357955] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.357965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.358094] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.358103] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.358199] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.358208] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.358411] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.358421] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.358596] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.358606] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.358694] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.358703] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.760 [2024-07-15 18:12:48.358852] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.760 [2024-07-15 18:12:48.358862] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.760 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.359034] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.359044] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.359165] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.359176] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.359349] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.359359] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.359512] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.359522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.359742] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.359771] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.359973] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.360003] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.360272] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.360282] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.360403] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.360413] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.360587] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.360597] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.360700] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.360709] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.360956] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.360986] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.361244] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.361274] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.361557] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.361567] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.361748] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.361758] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.361998] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.362007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.362177] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.362189] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.362307] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.362318] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.362483] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.362493] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.362595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.362604] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.362777] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.362787] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.362946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.362956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.363054] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.363063] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.363222] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.363236] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.363476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.363485] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.363588] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.363598] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.363867] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.363877] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.364108] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.364118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.364243] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.364253] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.364360] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.364369] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.364491] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.364501] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.364628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.364638] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.364808] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.364818] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.364942] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.364952] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.365053] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.365063] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.365169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.365179] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.365300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.365310] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.365484] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.365494] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.365658] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.365667] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.365864] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.365873] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.366047] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.366056] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.366170] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.366179] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.366419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.366429] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.366536] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.366546] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.366718] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.366727] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.366905] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.366914] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.367094] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.367104] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.367267] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.367277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.367429] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.367438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.367542] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.367552] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.367673] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.367682] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.367790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.367800] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.367917] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.367926] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.368095] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.368104] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.368266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.368276] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.368451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.368460] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.368525] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.368535] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.368638] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.368648] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.368818] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.368828] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.368952] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.761 [2024-07-15 18:12:48.368962] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.761 qpair failed and we were unable to recover it. 00:26:54.761 [2024-07-15 18:12:48.369127] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.369137] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.369232] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.369242] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.369403] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.369413] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.369592] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.369621] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.369773] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.369804] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.369962] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.369992] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.370190] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.370219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.370585] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.370594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.370858] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.370868] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.371047] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.371057] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.371170] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.371180] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.371453] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.371463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.371651] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.371661] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.371861] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.371871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.372069] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.372079] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.372181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.372191] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.372304] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.372315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.372474] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.372484] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.372647] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.372657] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.372743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.372752] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.372925] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.372935] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.373099] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.373133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.373246] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.373277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.373495] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.373525] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.373825] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.373835] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.374044] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.374053] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.374174] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.374184] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.374292] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.374302] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.374416] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.374426] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.374586] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.374596] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.374769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.374779] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.374958] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.374968] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.375159] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.375168] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.375287] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.375297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.375389] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.375398] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.375580] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.375589] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.375708] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.375720] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.375837] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.375847] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.375961] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.375971] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.376078] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.376088] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.376268] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.376278] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.376390] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.376400] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.376628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.376638] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.376736] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.376744] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.376840] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.376850] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.376958] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.376968] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.377139] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.377148] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.377249] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.377258] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.377359] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.377368] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.377536] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.377546] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.377786] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.377796] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.378029] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.378058] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.378251] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.378282] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.378621] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.378650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.378791] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.378820] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.379088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.379118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.379279] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.379289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.379468] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.379477] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.379658] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.379667] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.379824] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.379834] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.380028] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.380037] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.380215] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.380229] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.380411] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.380440] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.380697] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.380765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.380974] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.381041] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.381269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.381286] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.381486] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.381500] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.381773] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.762 [2024-07-15 18:12:48.381803] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.762 qpair failed and we were unable to recover it. 00:26:54.762 [2024-07-15 18:12:48.384522] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.384557] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.384876] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.384905] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.385123] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.385153] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.385378] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.385409] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.385674] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.385687] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.385945] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.385959] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.386156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.386170] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.386283] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.386297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.386426] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.386444] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.386616] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.386630] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.386744] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.386757] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.386936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.386965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.387165] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.387195] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.387446] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.387460] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.387698] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.387729] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.387882] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.387911] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.388045] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.388075] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.388276] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.388289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.388429] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.388444] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.388713] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.388742] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.388890] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.388920] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.389120] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.389150] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.389363] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.389394] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.389532] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.389561] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.389762] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.389776] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.389948] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.389962] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.390061] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.390074] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.390258] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.390272] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.390405] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.390437] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.390744] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.390774] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.391088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.391118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.391361] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.391392] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.391599] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.391613] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.391787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.391801] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.391932] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.391946] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.392178] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.392213] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.392427] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.392443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.392621] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.392634] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.392755] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.392768] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.392881] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.392894] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.393064] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.393078] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.393181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.393194] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.393299] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.393312] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.393545] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.393558] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.393732] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.393747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.393929] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.393958] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.394163] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.394193] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.394325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.394356] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.394594] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.394632] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.394849] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.394878] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.395152] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.395182] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.395412] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.395426] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.395604] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.395618] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.395792] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.395821] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.395966] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.395996] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.396293] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.396334] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.396420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.396433] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.396633] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.396646] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.396816] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.396829] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.396949] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.396962] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.397132] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.397146] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.397312] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.763 [2024-07-15 18:12:48.397325] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.763 qpair failed and we were unable to recover it. 00:26:54.763 [2024-07-15 18:12:48.397516] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.397546] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.397702] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.397731] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.397870] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.397900] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.398196] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.398232] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.398349] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.398362] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.398476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.398489] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.398672] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.398685] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.398796] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.398810] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.398936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.398963] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.399156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.399168] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.399276] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.399285] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.399401] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.399412] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.399580] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.399590] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.399718] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.399753] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.399883] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.399897] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.400030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.400043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.400221] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.400240] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.400358] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.400372] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.400542] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.400555] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.400792] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.400806] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.400998] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.401011] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.401116] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.401129] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.401249] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.401263] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.401392] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.401405] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.401577] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.401591] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.401776] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.401790] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.401915] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.401929] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.402042] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.402056] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.402312] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.402324] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.402436] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.402446] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.402609] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.402619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.402772] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.402782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.403019] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.403029] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.403130] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.403140] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.403263] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.403273] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.403433] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.403442] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.403620] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.403650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.403886] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.403916] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.404132] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.404174] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.404377] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.404387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.404565] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.404575] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.404688] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.404697] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.404923] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.404933] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.405161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.405171] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.405349] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.405360] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.405572] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.405602] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.405898] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.405927] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.406089] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.406118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.406356] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.406387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.406653] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.406663] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.406888] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.406897] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.407016] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.407026] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.407201] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.407211] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.407468] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.407480] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.407599] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.407610] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.407810] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.407820] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.407991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.408001] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.408108] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.408117] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.408296] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.408312] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.408436] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.408446] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.408615] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.408625] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.408772] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.408782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.408896] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.408906] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.409081] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.409091] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.409212] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.409222] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.409466] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.409476] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.409590] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.409600] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.409721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.409731] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.764 qpair failed and we were unable to recover it. 00:26:54.764 [2024-07-15 18:12:48.409822] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.764 [2024-07-15 18:12:48.409831] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.410008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.410018] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.410219] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.410266] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.410409] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.410439] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.410597] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.410626] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.410774] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.410803] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.411122] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.411151] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.411368] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.411398] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.411600] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.411610] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.411714] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.411723] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.411982] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.411992] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.412086] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.412095] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.412268] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.412291] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.412511] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.412540] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.412828] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.412857] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.413077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.413106] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.413332] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.413362] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.413626] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.413636] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.413863] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.413873] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.414069] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.414079] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.414178] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.414188] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.414357] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.414368] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.414563] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.414592] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.414795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.414825] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.414968] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.414997] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.415259] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.415295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.415508] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.415518] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.415635] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.415645] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.415806] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.415816] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.416015] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.416045] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.416154] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.416183] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.416391] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.416422] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.416625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.416635] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.416714] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.416722] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.416901] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.416910] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.417141] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.417150] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.417256] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.417266] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.417451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.417461] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.417552] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.417562] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.417684] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.417694] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.417798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.417808] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.417885] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.417894] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.418065] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.418074] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.418326] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.418337] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.418440] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.418450] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.418563] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.418572] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.418683] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.418692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.418817] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.418827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.419071] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.419081] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.419185] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.419194] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.419295] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.419304] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.419532] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.419541] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.419710] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.419720] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.419898] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.419908] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.420084] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.420114] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.420350] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.420381] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.420613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.420643] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.420864] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.420894] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.421179] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.421209] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.421499] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.421508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.421680] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.421689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.421861] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.421871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.421965] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.421974] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.422232] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.422242] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.422360] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.422369] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.422489] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.422501] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.422613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.422622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.422787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.422797] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.422932] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.422942] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.423100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.423110] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.423192] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.765 [2024-07-15 18:12:48.423201] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.765 qpair failed and we were unable to recover it. 00:26:54.765 [2024-07-15 18:12:48.423362] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.423372] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.423546] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.423556] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.423712] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.423722] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.423888] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.423898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.424074] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.424083] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.424258] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.424268] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.424368] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.424377] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.424566] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.424576] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.424746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.424756] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.424851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.424860] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.425056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.425066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.425241] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.425251] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.425482] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.425512] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.425661] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.425690] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.425858] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.425888] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.426181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.426211] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.426396] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.426427] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.426593] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.426634] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.426860] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.426870] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.426980] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.426990] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.427148] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.427158] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.427386] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.427396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.427502] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.427512] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.427689] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.427699] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.427889] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.427899] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.428001] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.428014] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.428130] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.428140] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.428260] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.428270] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.428439] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.428448] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.428557] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.428567] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.428667] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.428677] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.428762] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.428771] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.428883] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.428893] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.429007] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.429017] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.429197] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.429208] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.429279] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.429289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.429487] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.429496] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.429590] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.429599] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.429729] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.429739] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.429855] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.429865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.430037] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.430048] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.430227] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.430237] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.430483] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.430494] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.430669] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.430678] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.430859] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.430869] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.431052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.431063] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.431176] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.431186] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.431354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.431364] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.431483] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.431493] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.431651] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.431661] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.431865] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.431895] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.432114] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.432143] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.432430] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.432461] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.432666] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.432695] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.432860] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.432889] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.433186] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.433216] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.433354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.433384] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.433599] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.433630] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.433869] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.433899] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.434100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.434130] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.434368] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.434398] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.434548] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.434558] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.766 [2024-07-15 18:12:48.434719] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.766 [2024-07-15 18:12:48.434728] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.766 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.434823] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.434832] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.435004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.435015] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.435140] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.435150] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.435381] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.435412] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.435561] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.435591] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.435894] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.435924] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.436077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.436106] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.436307] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.436337] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.436471] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.436481] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.436657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.436666] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.436772] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.436781] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.436941] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.436955] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.437061] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.437070] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.437232] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.437242] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.437382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.437392] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.437597] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.437607] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.437778] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.437788] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.437911] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.437922] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.438028] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.438038] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.438152] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.438162] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.438270] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.438280] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.438380] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.438390] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.438573] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.438583] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.438762] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.438772] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.438891] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.438901] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.439019] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.439029] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.439119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.439129] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.439303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.439314] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.439428] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.439438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.439620] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.439630] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.439750] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.439760] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.439867] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.439877] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.440044] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.440054] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.440207] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.440217] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.440321] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.440331] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.440498] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.440508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.440631] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.440640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.440880] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.440909] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.441053] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.441083] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.441292] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.441324] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.441524] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.441534] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.441715] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.441725] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.441917] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.441947] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.442268] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.442299] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.442453] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.442483] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.442645] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.442674] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.442882] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.442912] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.443065] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.443094] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.443387] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.443417] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.443571] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.443582] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.443744] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.443754] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.443828] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.443839] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.444074] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.444084] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:54.767 [2024-07-15 18:12:48.444328] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:54.767 [2024-07-15 18:12:48.444338] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:54.767 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.444423] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.444432] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.444535] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.444544] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.444710] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.444722] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.444872] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.444882] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.444984] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.444993] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.445136] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.445146] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.445280] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.445289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.445406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.445415] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.445506] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.445515] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.445642] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.445652] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.445770] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.445778] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.445959] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.445968] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.446095] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.446104] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.446210] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.446219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.446370] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.446380] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.446486] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.446496] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.446594] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.446603] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.446696] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.446705] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.446817] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.446826] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.446927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.446936] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.447035] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.447044] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.447219] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.447232] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.447355] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.447365] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.447531] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.447541] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.447654] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.447664] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.447862] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.447872] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.448051] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.448062] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.448243] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.448273] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.448433] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.448463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.448698] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.448727] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.448999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.449009] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.449212] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.449223] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.449394] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.060 [2024-07-15 18:12:48.449404] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.060 qpair failed and we were unable to recover it. 00:26:55.060 [2024-07-15 18:12:48.449519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.449529] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.449630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.449640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.449818] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.449827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.449943] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.449953] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.450129] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.450141] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.450257] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.450268] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.450360] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.450372] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.450622] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.450632] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.450807] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.450818] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.450920] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.450930] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.451096] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.451106] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.451216] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.451230] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.451397] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.451407] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.451577] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.451587] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.451822] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.451852] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.452014] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.452042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.452181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.452210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.452346] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.452376] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.452591] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.452620] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.452865] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.452894] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.453206] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.453243] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.454382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.454404] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.454667] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.454678] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.454906] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.454916] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.455123] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.455132] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.455263] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.455273] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.455383] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.455394] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.455517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.061 [2024-07-15 18:12:48.455526] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.061 qpair failed and we were unable to recover it. 00:26:55.061 [2024-07-15 18:12:48.455676] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.455705] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.455866] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.455895] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.456099] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.456128] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.456302] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.456347] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.456577] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.456608] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.456811] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.456825] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.456979] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.457010] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.457242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.457272] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.457490] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.457520] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.457811] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.457840] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.459015] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.459040] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.459347] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.459362] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.459542] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.459556] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.459769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.459799] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.460119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.460148] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.460317] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.460347] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.460587] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.460615] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.460907] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.460922] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.461049] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.461080] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.461237] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.461269] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.461470] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.461499] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.461709] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.461722] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.461863] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.461876] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.462044] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.462058] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.462187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.462201] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.462331] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.462345] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.462554] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.462568] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.462665] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.462678] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.462860] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.462874] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.463017] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.463031] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.463206] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.463219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.463466] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.463477] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.463690] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.463699] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.463806] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.463816] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.463922] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.463931] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.464032] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.464041] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.464205] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.464214] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.464461] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.464472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.464637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.464648] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.464754] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.062 [2024-07-15 18:12:48.464764] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.062 qpair failed and we were unable to recover it. 00:26:55.062 [2024-07-15 18:12:48.464876] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.464886] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.465068] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.465078] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.465190] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.465199] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.465307] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.465320] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.465487] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.465497] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.465600] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.465610] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.465770] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.465780] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.465946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.465956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.466065] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.466076] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.466176] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.466186] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.466422] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.466432] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.466544] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.466554] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.466640] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.466649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.466833] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.466844] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.466949] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.466961] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.467065] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.467076] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.467155] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.467167] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.467272] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.467282] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.467442] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.467452] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.467566] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.467576] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.467763] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.467772] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.467859] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.467868] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.468004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.468014] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.468121] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.468131] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.468363] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.468374] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.468474] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.468483] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.468602] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.468612] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.468791] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.468801] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.468993] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.469004] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.469161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.469171] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.469347] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.469360] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.469475] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.469485] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.469595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.469605] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.469765] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.469775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.469944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.469954] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.470106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.470117] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.470290] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.470300] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.470418] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.470430] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.470530] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.470541] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.470645] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.470656] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.063 [2024-07-15 18:12:48.470837] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.063 [2024-07-15 18:12:48.470846] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.063 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.470940] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.470950] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.471112] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.471122] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.471302] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.471312] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.471418] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.471429] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.471548] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.471558] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.471727] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.471739] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.471847] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.471857] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.471962] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.471971] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.472136] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.472146] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.472261] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.472270] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.472385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.472395] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.472509] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.472519] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.472611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.472619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.472710] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.472718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.472895] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.472905] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.473009] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.473021] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.473097] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.473107] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.473216] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.473229] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.473350] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.473360] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.473589] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.473599] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.473693] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.473702] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.473814] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.473824] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.473990] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.474000] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.474097] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.474109] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.474229] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.474239] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.474354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.474364] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.474457] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.474466] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.474646] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.474674] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.474817] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.474846] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.475062] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.475096] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.475246] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.475276] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.064 [2024-07-15 18:12:48.475437] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.064 [2024-07-15 18:12:48.475466] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.064 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.475610] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.475639] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.475875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.475904] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.476243] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.476273] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.476425] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.476456] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.476608] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.476618] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.476786] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.476795] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.476895] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.476907] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.477021] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.477032] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.477188] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.477198] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.477394] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.477405] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.477517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.477530] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.477694] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.477704] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.477818] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.477827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.478001] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.478011] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.478175] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.478185] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.478295] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.478304] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.478401] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.478410] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.478506] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.478515] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.478684] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.478694] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.478914] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.478943] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.479163] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.479192] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.479353] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.479384] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.479582] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.479592] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.479704] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.479714] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.479817] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.479827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.480052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.480062] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.480216] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.480255] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.480495] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.480526] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.480682] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.480711] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.480852] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.480861] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.480962] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.480971] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.481222] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.481237] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.481448] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.481477] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.481680] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.481709] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.481923] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.481952] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.482120] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.482150] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.482362] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.482392] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.482550] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.482585] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.482803] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.482832] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.065 [2024-07-15 18:12:48.482983] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.065 [2024-07-15 18:12:48.483013] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.065 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.483161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.483189] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.483349] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.483379] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.483591] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.483621] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.483762] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.483772] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.483865] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.483875] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.483984] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.483993] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.484098] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.484107] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.484275] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.484285] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.484446] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.484456] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.484557] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.484567] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.484730] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.484740] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.484898] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.484908] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.485070] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.485080] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.485193] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.485203] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.485332] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.485342] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.485525] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.485534] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.485718] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.485729] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.485855] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.485865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.485967] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.485977] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.486145] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.486154] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.486258] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.486268] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.486386] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.486396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.486486] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.486498] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.486604] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.486614] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.486711] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.486721] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.486833] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.486843] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.486940] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.486950] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.487109] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.487119] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.487372] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.487384] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.487545] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.487557] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.487651] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.487660] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.487760] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.487770] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.487864] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.487874] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.488022] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.488032] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.488148] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.488158] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.488285] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.488295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.488389] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.488399] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.488595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.488611] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.488774] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.488784] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.066 [2024-07-15 18:12:48.488981] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.066 [2024-07-15 18:12:48.488991] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.066 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.489131] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.489142] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.489259] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.489269] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.489429] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.489438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.489599] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.489609] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.489707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.489717] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.489941] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.489951] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.490045] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.490054] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.490169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.490179] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.490357] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.490367] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.490547] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.490557] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.490663] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.490674] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.490837] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.490848] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.491008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.491018] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.491140] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.491150] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.491266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.491276] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.491392] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.491402] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.491519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.491529] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.491630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.491640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.491735] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.491745] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.491913] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.491923] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.492082] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.492092] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.492161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.492171] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.492338] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.492348] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.492449] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.492479] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.492624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.492653] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.492873] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.492903] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.493106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.493116] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.493232] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.493242] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.493494] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.493504] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.493663] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.493673] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.493786] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.493796] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.493973] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.493983] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.494079] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.494089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.494184] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.494194] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.494294] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.494304] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.494467] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.494477] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.494675] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.494685] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.494812] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.494823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.494996] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.495026] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.495179] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.495209] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.067 [2024-07-15 18:12:48.495442] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.067 [2024-07-15 18:12:48.495471] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.067 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.495603] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.495613] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.495722] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.495732] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.495822] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.495832] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.495944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.495954] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.496132] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.496142] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.496302] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.496312] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.496475] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.496484] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.496589] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.496599] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.496701] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.496711] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.496815] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.496825] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.497002] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.497012] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.497172] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.497181] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.497277] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.497287] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.497394] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.497404] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.497512] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.497522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.497645] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.497655] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.497766] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.497776] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.497955] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.497965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.498188] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.498217] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.498345] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.498375] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.498595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.498605] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.498723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.498733] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.498891] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.498900] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.499080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.499090] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.499202] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.499212] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.499451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.499461] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.499621] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.499631] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.499733] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.499743] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.499842] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.499852] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.499960] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.499970] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.500154] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.500164] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.500294] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.500305] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.500423] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.500433] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.500695] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.500706] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.500934] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.500944] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.501053] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.501063] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.501157] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.501167] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.501346] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.501357] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.501455] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.501463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.501637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.501647] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.068 qpair failed and we were unable to recover it. 00:26:55.068 [2024-07-15 18:12:48.501769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.068 [2024-07-15 18:12:48.501779] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.501882] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.501891] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.502071] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.502082] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.502257] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.502268] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.502372] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.502381] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.502487] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.502497] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.502656] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.502666] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.502861] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.502890] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.503109] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.503138] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.503255] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.503285] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.503441] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.503471] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.503747] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.503776] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.503973] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.503983] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.504236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.504246] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.504366] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.504375] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.504555] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.504565] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.504677] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.504687] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.504854] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.504892] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.505041] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.505070] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.505208] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.505244] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.505402] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.505431] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.505665] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.505675] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.505846] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.505856] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.505991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.506001] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.506128] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.506138] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.506248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.506258] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.506352] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.506361] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.506487] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.506497] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.506621] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.506631] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.506730] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.506739] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.506834] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.506844] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.506946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.506955] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.507123] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.507133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.507234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.507244] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.507342] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.069 [2024-07-15 18:12:48.507351] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.069 qpair failed and we were unable to recover it. 00:26:55.069 [2024-07-15 18:12:48.507431] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.507440] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.507543] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.507555] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.507702] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.507712] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.507851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.507861] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.507968] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.507978] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.508174] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.508184] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.508356] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.508367] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.508458] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.508467] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.508590] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.508600] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.508759] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.508769] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.508882] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.508892] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.508996] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.509006] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.509123] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.509134] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.509242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.509252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.509418] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.509428] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.509513] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.509522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.509621] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.509630] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.509739] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.509749] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.509918] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.509927] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.510051] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.510060] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.510286] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.510317] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.510454] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.510484] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.510695] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.510724] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.510846] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.510856] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.510969] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.510979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.511147] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.511157] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.511331] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.511341] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.511572] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.511582] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.511719] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.511729] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.511911] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.511921] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.070 [2024-07-15 18:12:48.512031] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.070 [2024-07-15 18:12:48.512041] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.070 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.512223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.512262] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.512482] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.512511] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.512654] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.512683] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.512813] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.512823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.512936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.512946] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.513051] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.513061] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.513161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.513171] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.513398] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.513408] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.513574] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.513583] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.513747] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.513756] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.513866] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.513877] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.514040] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.514049] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.514232] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.514242] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.514350] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.514359] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.514447] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.514456] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.514628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.514638] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.514808] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.514837] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.515042] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.515071] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.515203] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.515240] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.515454] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.515484] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.515718] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.515747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.515890] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.515929] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.516114] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.516124] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.516292] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.516302] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.516478] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.516517] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.516660] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.516689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.516853] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.516882] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.517021] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.517050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.517263] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.517293] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.517426] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.517455] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.517657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.517666] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.517764] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.517774] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.517941] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.517951] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.518123] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.518133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.518306] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.518317] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.518437] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.518447] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.518623] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.518633] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.071 [2024-07-15 18:12:48.518759] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.071 [2024-07-15 18:12:48.518769] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.071 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.518867] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.518877] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.518996] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.519006] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.519185] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.519195] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.519308] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.519318] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.519426] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.519436] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.519535] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.519544] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.519735] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.519745] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.519854] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.519864] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.519960] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.519970] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.520082] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.520093] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.520273] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.520283] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.520399] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.520409] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.520599] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.520612] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.520734] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.520743] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.520834] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.520844] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.520938] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.520948] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.521049] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.521059] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.521149] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.521158] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.521261] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.521271] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.521449] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.521459] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.521636] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.521645] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.521737] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.521747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.521849] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.521858] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.521971] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.521981] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.522094] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.522105] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.522300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.522311] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.522530] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.522560] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.522697] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.522726] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.522928] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.522957] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.523099] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.523128] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.523336] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.072 [2024-07-15 18:12:48.523366] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.072 qpair failed and we were unable to recover it. 00:26:55.072 [2024-07-15 18:12:48.523510] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.523539] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.523746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.523775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.524007] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.524036] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.524187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.524216] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.524446] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.524475] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.524608] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.524617] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.524789] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.524799] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.524977] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.524986] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.525148] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.525215] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.525456] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.525491] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.525648] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.525678] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.525898] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.525911] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.526084] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.526097] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.526194] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.526208] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.526400] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.526415] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.526522] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.526535] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.526738] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.526752] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.526877] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.526892] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.527053] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.527067] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.527242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.527257] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.527362] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.527376] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.527491] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.527504] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.527627] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.527641] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.527742] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.527756] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.527928] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.527941] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.528056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.528069] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.528219] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.528237] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.528344] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.528358] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.528478] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.528492] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.528618] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.528630] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.528810] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.528839] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.528958] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.528988] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.529302] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.529332] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.529495] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.529524] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.529666] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.529675] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.529887] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.529919] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.530195] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.073 [2024-07-15 18:12:48.530235] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.073 qpair failed and we were unable to recover it. 00:26:55.073 [2024-07-15 18:12:48.530443] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.530472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.530637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.530667] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.530903] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.530933] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.531154] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.531167] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.531278] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.531292] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.531409] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.531422] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.531681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.531695] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.531882] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.531896] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.532023] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.532037] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.532242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.532271] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.532495] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.532524] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.532730] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.532760] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.532922] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.532936] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.533206] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.533243] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.533406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.533438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.533570] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.533599] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.533749] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.533762] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.533959] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.533989] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.534151] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.534181] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.534406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.534436] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.534582] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.534612] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.534850] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.534887] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.535056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.535070] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.535187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.535200] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.535438] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.535469] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.535630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.535663] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.535898] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.535911] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.536034] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.536048] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.536206] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.536219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.536410] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.536420] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.536600] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.536610] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.536727] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.536736] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.536911] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.536922] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.537114] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.537144] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.537361] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.537392] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.537625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.074 [2024-07-15 18:12:48.537667] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.074 qpair failed and we were unable to recover it. 00:26:55.074 [2024-07-15 18:12:48.537854] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.537864] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.538041] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.538071] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.538211] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.538249] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.538478] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.538507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.538660] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.538689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.538823] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.538852] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.539009] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.539043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.539221] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.539234] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.539393] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.539404] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.539518] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.539528] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.539646] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.539655] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.539821] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.539831] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.540008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.540018] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.540120] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.540130] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.540300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.540311] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.540421] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.540431] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.540613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.540624] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.540791] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.540801] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.540925] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.540935] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.541042] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.541052] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.541167] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.541176] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.541354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.541365] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.541559] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.541569] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.541675] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.541685] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.541809] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.541819] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.541914] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.541923] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.542029] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.542039] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.542217] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.075 [2024-07-15 18:12:48.542231] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.075 qpair failed and we were unable to recover it. 00:26:55.075 [2024-07-15 18:12:48.542357] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.542367] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.542465] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.542477] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.542583] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.542593] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.542686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.542695] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.542804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.542814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.542922] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.542932] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.543093] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.543103] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.543193] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.543202] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.543401] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.543411] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.543640] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.543650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.543772] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.543781] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.543865] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.543874] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.543985] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.543995] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.544098] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.544108] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.544266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.544275] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.544504] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.544514] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.544623] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.544633] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.544728] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.544738] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.544981] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.544991] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.545141] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.545151] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.545354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.545365] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.545476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.545486] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.545591] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.545601] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.545728] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.545738] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.545835] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.545844] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.546028] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.546038] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.546215] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.546230] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.076 qpair failed and we were unable to recover it. 00:26:55.076 [2024-07-15 18:12:48.546330] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.076 [2024-07-15 18:12:48.546339] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.546515] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.546525] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.546717] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.546746] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.546886] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.546915] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.547133] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.547162] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.547301] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.547332] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.547558] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.547587] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.547790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.547819] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.547991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.548020] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.548194] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.548223] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.548381] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.548411] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.548612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.548622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.548827] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.548857] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.549010] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.549040] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.549185] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.549219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.549434] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.549464] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.549641] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.549670] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.549831] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.549860] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.549998] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.550027] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.550240] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.550250] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.550406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.550416] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.550573] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.550602] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.550809] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.550838] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.550992] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.551022] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.551177] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.551205] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.551379] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.551409] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.551560] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.551589] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.551792] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.551802] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.077 qpair failed and we were unable to recover it. 00:26:55.077 [2024-07-15 18:12:48.551971] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.077 [2024-07-15 18:12:48.552000] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.552140] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.552169] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.552333] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.552364] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.552566] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.552595] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.552752] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.552762] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.552867] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.552877] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.553015] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.553025] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.553199] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.553241] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.553398] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.553426] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.553568] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.553597] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.553862] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.553872] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.554056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.554075] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.554199] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.554210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.554359] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.554370] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.554483] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.554493] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.554662] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.554672] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.554782] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.554792] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.554916] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.554926] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.555044] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.555053] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.555214] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.555228] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.555340] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.555349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.555518] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.555528] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.555637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.555647] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.555740] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.555749] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.555923] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.555934] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.556041] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.556051] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.556144] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.556155] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.556267] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.078 [2024-07-15 18:12:48.556284] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.078 qpair failed and we were unable to recover it. 00:26:55.078 [2024-07-15 18:12:48.556459] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.556468] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.556561] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.556573] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.556665] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.556675] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.556778] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.556788] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.556886] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.556896] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.557003] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.557013] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.557118] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.557127] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.557255] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.557265] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.557358] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.557367] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.557462] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.557472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.557565] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.557575] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.557732] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.557741] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.557830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.557840] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.557936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.557946] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.558203] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.558213] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.558324] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.558334] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.558519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.558529] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.558637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.558647] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.558827] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.558857] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.559002] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.559030] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.559262] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.559292] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.559445] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.559475] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.559680] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.079 [2024-07-15 18:12:48.559709] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.079 qpair failed and we were unable to recover it. 00:26:55.079 [2024-07-15 18:12:48.559846] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.559875] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.560012] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.560050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.560162] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.560172] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.560348] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.560358] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.560457] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.560467] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.560569] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.560579] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.560749] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.560759] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.560883] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.560894] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.561006] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.561016] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.561139] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.561149] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.561260] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.561270] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.561364] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.561374] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.561551] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.561561] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.561655] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.561665] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.561756] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.561766] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.561873] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.561886] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.561986] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.561996] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.562085] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.562095] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.562187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.562196] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.562377] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.562387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.562496] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.562505] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.562685] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.562694] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.562863] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.562873] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.562970] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.562981] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.563163] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.563173] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.563279] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.563289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.563406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.563415] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.563512] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.563522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.563627] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.563637] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.563814] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.563824] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.563936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.563946] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.564047] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.564056] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.564167] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.564177] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.564341] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.564351] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.564456] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.564466] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.080 [2024-07-15 18:12:48.564589] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.080 [2024-07-15 18:12:48.564599] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.080 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.564694] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.564704] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.564914] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.564923] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.565033] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.565043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.565142] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.565152] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.565248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.565258] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.565421] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.565431] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.565591] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.565601] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.565692] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.565702] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.565808] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.565818] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.565912] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.565922] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.566019] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.566029] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.566189] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.566198] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.566358] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.566367] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.566528] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.566538] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.566648] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.566657] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.566765] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.566776] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.566888] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.566898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.567003] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.567013] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.567231] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.567241] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.567337] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.567349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.567451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.567461] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.567572] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.567581] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.567675] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.567685] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.567854] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.567864] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.567973] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.567982] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.568081] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.568091] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.568185] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.568195] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.568297] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.568308] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.568469] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.568479] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.568639] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.568648] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.568746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.568756] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.568860] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.568871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.568967] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.568976] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.569070] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.569080] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.569343] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.569353] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.569461] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.569470] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.081 [2024-07-15 18:12:48.569615] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.081 [2024-07-15 18:12:48.569625] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.081 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.569789] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.569799] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.569976] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.570006] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.570156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.570186] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.570336] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.570366] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.570665] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.570694] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.570819] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.570829] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.570935] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.570945] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.571041] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.571050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.571169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.571179] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.571435] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.571445] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.571537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.571547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.571667] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.571677] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.571773] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.571782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.571874] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.571883] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.572057] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.572067] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.572177] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.572187] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.572416] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.572427] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.572520] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.572530] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.572698] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.572708] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.572949] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.572959] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.573064] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.573073] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.573177] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.573187] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.573304] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.573316] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.573423] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.573432] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.573595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.573604] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.573702] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.573711] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.573808] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.573817] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.573955] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.573965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.574075] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.574084] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.574187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.574197] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.574309] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.574319] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.574428] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.574437] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.574541] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.082 [2024-07-15 18:12:48.574551] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.082 qpair failed and we were unable to recover it. 00:26:55.082 [2024-07-15 18:12:48.574652] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.574662] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.574773] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.574782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.574972] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.574982] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.575143] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.575153] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.575262] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.575272] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.575455] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.575465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.575584] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.575594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.575846] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.575857] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.576025] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.576035] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.576203] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.576244] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.576455] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.576485] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.576757] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.576786] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.576927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.576956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.577172] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.577202] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.577433] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.577464] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.577607] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.577636] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.578011] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.578079] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.578244] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.578279] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.578500] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.578531] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.578822] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.578852] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.578987] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.579001] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.579120] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.579133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.579306] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.579320] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.579511] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.579541] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.579689] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.579720] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.579919] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.579948] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.580108] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.580121] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.580219] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.580237] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.580408] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.580421] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.580541] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.580555] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.580674] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.580687] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.580783] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.580796] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.580981] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.580995] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.581102] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.581115] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.083 qpair failed and we were unable to recover it. 00:26:55.083 [2024-07-15 18:12:48.581220] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.083 [2024-07-15 18:12:48.581240] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.581465] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.581479] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.581637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.581650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.581757] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.581771] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.581962] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.581991] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.582142] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.582171] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.582345] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.582375] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.582529] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.582559] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.582705] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.582735] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.582962] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.583009] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.583204] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.583218] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.583346] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.583359] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.583468] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.583482] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.583655] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.583685] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.583875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.583904] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.584051] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.584081] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.584314] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.584327] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.584439] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.584452] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.584626] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.584640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.584833] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.584846] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.584967] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.584981] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.585199] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.585235] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.585459] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.585489] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.585649] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.585679] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.585880] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.585893] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.586100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.586130] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.586282] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.586312] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.586467] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.586496] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.586647] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.084 [2024-07-15 18:12:48.586677] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.084 qpair failed and we were unable to recover it. 00:26:55.084 [2024-07-15 18:12:48.586810] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.586823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.586922] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.586935] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.587143] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.587156] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.587414] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.587427] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.587598] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.587612] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.587783] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.587797] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.587912] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.587925] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.588143] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.588159] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.588279] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.588293] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.588415] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.588428] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.588562] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.588576] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.588745] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.588759] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.588946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.588960] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.589061] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.589074] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.589240] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.589254] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.589367] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.589380] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.589484] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.589498] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.589699] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.589712] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.589822] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.589836] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.589948] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.589961] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.590073] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.590086] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.590218] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.590237] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.590411] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.590425] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.590526] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.590539] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.590682] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.590712] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.590861] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.590891] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.591030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.591060] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.591188] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.591202] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.591325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.591339] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.591507] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.591521] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.591723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.591752] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.591889] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.591919] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.592071] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.592101] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.592310] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.592341] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.592481] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.592516] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.592784] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.592798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.592969] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.592983] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.593110] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.593124] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.593245] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.593259] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.593442] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.593471] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.085 [2024-07-15 18:12:48.593674] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.085 [2024-07-15 18:12:48.593703] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.085 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.593927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.593968] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.594073] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.594086] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.594274] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.594288] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.594394] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.594408] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.594603] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.594616] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.594721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.594735] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.594843] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.594856] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.595011] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.595045] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.595223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.595239] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.595400] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.595410] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.595509] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.595519] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.595624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.595634] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.595733] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.595743] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.595844] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.595854] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.595955] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.595965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.596075] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.596085] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.596232] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.596242] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.596414] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.596425] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.596602] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.596612] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.596722] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.596732] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.596828] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.596839] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.596936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.596945] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.597110] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.597120] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.597231] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.597241] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.597362] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.597372] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.597547] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.597557] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.597686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.597696] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.597805] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.597815] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.597999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.598009] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.598106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.598116] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.598219] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.598234] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.598339] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.598349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.598452] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.598461] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.598576] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.598586] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.598691] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.598701] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.598803] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.598812] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.598913] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.598923] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.599025] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.599035] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.599195] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.599205] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.599320] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.599330] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.086 [2024-07-15 18:12:48.599496] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.086 [2024-07-15 18:12:48.599505] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.086 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.599626] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.599635] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.599803] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.599813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.599923] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.599933] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.600031] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.600040] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.600147] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.600157] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.600266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.600276] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.600400] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.600424] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.600611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.600624] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.600754] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.600768] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.600937] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.600951] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.601077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.601090] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.601204] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.601218] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.601347] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.601358] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.601549] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.601558] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.601655] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.601665] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.601762] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.601772] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.601869] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.601878] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.601974] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.601983] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.602096] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.602106] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.602201] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.602213] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.602315] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.602325] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.602430] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.602440] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.602533] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.602543] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.602654] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.602663] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.602905] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.602916] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.603086] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.603096] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.603282] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.603293] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.603399] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.603410] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.603504] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.603514] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.603679] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.603705] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.603856] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.603885] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.604031] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.604060] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.604255] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.604285] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.604503] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.604533] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.604746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.604776] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.604988] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.604997] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.605103] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.605113] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.605269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.605299] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.605519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.605547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.605692] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.605721] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.087 [2024-07-15 18:12:48.605867] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.087 [2024-07-15 18:12:48.605876] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.087 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.606042] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.606051] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.606219] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.606233] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.606344] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.606353] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.606516] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.606526] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.606626] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.606636] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.606816] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.606826] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.606992] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.607021] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.607170] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.607199] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.607479] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.607547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.607719] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.607752] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.607999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.608029] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.608170] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.608184] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.608302] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.608316] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.608511] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.608525] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.608697] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.608711] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.608916] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.608946] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.609170] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.609198] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.609443] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.609476] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.609750] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.609789] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.610001] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.610030] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.610239] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.610268] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.610422] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.610451] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.610667] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.610696] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.610907] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.610917] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.611018] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.611028] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.611268] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.611279] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.611446] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.611455] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.611638] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.611668] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.611879] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.611908] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.612130] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.612159] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.612368] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.612379] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.612483] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.612493] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.612586] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.612595] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.612698] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.612707] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.612807] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.612817] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.612923] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.612933] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.613048] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.613058] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.613156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.613165] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.613351] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.613361] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.613556] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.613566] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.088 qpair failed and we were unable to recover it. 00:26:55.088 [2024-07-15 18:12:48.613677] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.088 [2024-07-15 18:12:48.613687] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.613793] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.613803] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.614073] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.614082] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.614253] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.614263] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.614372] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.614381] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.614488] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.614498] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.614613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.614623] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.614784] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.614793] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.614954] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.614963] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.615049] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.615059] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.615305] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.615315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.615482] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.615493] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.615604] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.615613] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.615705] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.615714] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.615804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.615813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.615930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.615939] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.616036] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.616045] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.616165] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.616175] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.616373] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.616385] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.616546] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.616556] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.616677] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.616687] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.616926] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.616937] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.617100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.617120] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.617261] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.617272] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.617460] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.617470] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.617594] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.617604] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.617715] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.617725] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.617884] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.617925] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.618135] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.618165] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.618320] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.618350] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.618597] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.618625] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.618801] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.618830] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.619125] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.619157] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.619378] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.619410] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.619626] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.619656] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.089 qpair failed and we were unable to recover it. 00:26:55.089 [2024-07-15 18:12:48.619807] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.089 [2024-07-15 18:12:48.619837] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.619987] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.620018] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.620148] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.620160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.620281] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.620292] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.620411] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.620423] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.620598] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.620610] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.620775] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.620787] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.620950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.620962] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.621080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.621092] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.621211] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.621223] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.621333] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.621345] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.621501] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.621532] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.621709] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.621739] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.621884] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.621915] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.622056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.622088] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.622302] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.622333] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.622488] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.622518] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.622658] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.622707] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.622850] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.622882] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.623075] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.623105] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.623966] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.623987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.624162] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.624191] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.624348] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.624360] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.625645] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.625666] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.625887] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.625900] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.626060] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.626072] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.626697] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.626717] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.626909] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.626921] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.627143] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.627156] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.627284] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.627297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.627415] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.627427] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.627582] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.627594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.627690] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.627700] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.627809] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.627821] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.628021] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.628032] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.628187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.628198] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.628309] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.628320] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.628435] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.628446] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.628555] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.628586] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.628853] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.628864] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.629024] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.090 [2024-07-15 18:12:48.629067] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.090 qpair failed and we were unable to recover it. 00:26:55.090 [2024-07-15 18:12:48.629220] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.629272] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.629427] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.629458] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.629677] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.629707] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.629855] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.629886] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.630098] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.630130] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.630329] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.630341] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.630457] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.630467] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.630698] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.630710] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.630906] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.630937] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.631208] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.631254] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.631408] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.631439] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.631657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.631688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.631847] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.631877] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.632020] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.632051] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.632198] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.632239] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.632535] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.632578] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.632795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.632825] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.633042] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.633054] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.633223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.633241] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.633412] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.633443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.633620] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.633650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.633801] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.633831] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.633982] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.634012] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.634177] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.634209] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.634452] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.634483] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.634624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.634655] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.634813] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.634824] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.634930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.634943] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.635114] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.635126] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.635223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.635239] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.635345] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.635356] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.635466] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.635478] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.635636] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.635648] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.635740] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.635750] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.635858] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.635868] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.635987] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.635998] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.636113] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.636124] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.636255] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.636268] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.636432] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.636445] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.636616] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.636628] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.091 qpair failed and we were unable to recover it. 00:26:55.091 [2024-07-15 18:12:48.636793] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.091 [2024-07-15 18:12:48.636804] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.636912] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.636924] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.637037] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.637068] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.637213] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.637252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.637406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.637438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.637658] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.637689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.637861] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.637892] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.638046] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.638077] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.638289] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.638302] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.638531] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.638546] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.638713] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.638744] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.638957] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.638988] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.639202] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.639240] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.639515] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.639546] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.639687] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.639718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.639921] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.639933] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.640045] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.640056] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.640154] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.640164] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.640327] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.640340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.640439] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.640450] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.640556] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.640567] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.640683] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.640693] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.640813] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.640823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.641066] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.641078] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.641184] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.641194] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.641315] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.641326] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.641434] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.641447] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.641626] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.641638] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.641741] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.641751] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.641901] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.641913] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.642012] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.642022] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.642188] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.642200] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.642310] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.642321] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.642484] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.642496] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.642657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.642669] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.642837] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.642853] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.643027] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.643038] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.643156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.643167] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.643350] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.643362] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.643474] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.643486] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.643721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.092 [2024-07-15 18:12:48.643734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.092 qpair failed and we were unable to recover it. 00:26:55.092 [2024-07-15 18:12:48.643893] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.643905] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.644026] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.644039] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.644211] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.644223] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.644333] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.644345] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.644453] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.644464] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.644635] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.644647] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.644747] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.644758] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.644870] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.644882] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.645048] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.645062] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.645156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.645167] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.645348] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.645359] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.645471] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.645483] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.645591] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.645603] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.645773] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.645785] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.645939] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.645951] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.646137] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.646149] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.646311] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.646323] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.646487] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.646498] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.646611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.646622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.646719] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.646730] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.646835] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.646847] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.646961] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.646972] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.647142] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.647153] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.647326] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.647338] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.647580] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.647592] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.647762] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.647773] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.647893] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.647905] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.648092] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.648104] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.648202] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.648214] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.648319] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.648332] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.648453] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.648465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.648628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.648640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.648735] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.648747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.648864] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.648876] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.648975] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.648987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.649145] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.649156] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.649248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.649259] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.093 qpair failed and we were unable to recover it. 00:26:55.093 [2024-07-15 18:12:48.649372] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.093 [2024-07-15 18:12:48.649383] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.649560] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.649571] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.649733] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.649745] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.649837] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.649847] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.649955] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.649967] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.650127] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.650138] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.650316] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.650329] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.650435] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.650447] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.650585] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.650597] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.650706] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.650717] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.650875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.650887] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.651008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.651022] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.651134] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.651146] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.651315] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.651326] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.651436] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.651447] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.651612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.651623] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.651811] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.651822] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.652009] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.652021] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.652126] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.652138] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.652235] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.652246] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.652352] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.652364] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.652463] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.652474] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.652585] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.652597] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.652762] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.652774] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.652998] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.653009] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.653182] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.653195] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.653382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.653395] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.653563] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.653574] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.653733] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.653744] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.653870] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.653881] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.653973] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.653983] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.654144] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.654156] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.654394] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.654406] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.654516] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.654528] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.654699] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.654711] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.654805] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.654816] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.654918] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.654930] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.655040] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.655051] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.655282] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.655294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.655395] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.655407] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.655500] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.655512] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.655620] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.094 [2024-07-15 18:12:48.655631] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.094 qpair failed and we were unable to recover it. 00:26:55.094 [2024-07-15 18:12:48.655844] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.655856] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.655957] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.655969] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.656083] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.656096] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.656201] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.656232] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.656331] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.656343] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.656507] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.656518] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.656623] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.656635] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.656734] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.656746] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.656912] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.656925] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.657117] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.657130] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.657235] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.657246] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.657394] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.657405] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.657519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.657530] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.657636] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.657648] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.657923] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.657935] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.658034] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.658045] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.658222] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.658252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.658355] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.658366] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.658469] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.658481] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.658581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.658592] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.658703] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.658715] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.658834] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.658845] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.659016] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.659027] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.659200] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.659212] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.659311] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.659322] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.659435] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.659447] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.659537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.659548] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.659713] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.659724] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.659834] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.659845] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.660021] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.660032] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.660196] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.660207] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.660375] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.660388] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.660493] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.660505] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.660591] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.660601] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.660711] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.660722] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.660821] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.660834] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.660933] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.660945] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.661059] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.661071] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.661190] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.661201] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.661428] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.661441] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.661560] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.661571] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.661666] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.661678] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.661787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.661799] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.095 [2024-07-15 18:12:48.661966] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.095 [2024-07-15 18:12:48.661978] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.095 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.662141] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.662153] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.662294] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.662305] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.662415] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.662427] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.662563] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.662575] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.662690] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.662702] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.662871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.662884] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.663056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.663068] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.663194] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.663206] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.663385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.663397] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.663491] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.663503] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.663662] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.663673] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.663764] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.663775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.663956] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.663968] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.664143] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.664154] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.664265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.664277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.664391] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.664403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.664566] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.664577] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.664683] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.664694] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.664812] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.664824] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.664934] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.664946] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.665198] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.665210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.665335] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.665347] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.665446] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.665457] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.665556] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.665567] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.665689] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.665700] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.665839] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.665850] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.666019] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.666031] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.666249] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.666262] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.666390] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.666401] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.666526] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.666538] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.666665] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.666676] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.666777] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.666788] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.666889] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.666901] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.667000] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.667012] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.667177] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.667190] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.667305] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.667317] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.667403] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.667415] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.667534] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.667546] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.667651] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.667662] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.667826] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.667837] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.668000] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.668012] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.096 [2024-07-15 18:12:48.668171] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.096 [2024-07-15 18:12:48.668184] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.096 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.668282] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.668295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.668412] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.668423] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.668570] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.668582] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.668697] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.668710] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.668824] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.668835] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.669006] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.669018] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.669270] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.669282] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.669385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.669397] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.669549] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.669561] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.669684] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.669696] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.669811] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.669822] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.669924] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.669936] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.670032] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.670043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.670137] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.670149] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.670269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.670281] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.670385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.670397] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.670584] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.670596] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.670734] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.670746] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.670928] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.670940] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.671172] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.671183] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.671360] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.671373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.671562] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.671573] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.671678] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.671690] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.671868] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.671879] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.671978] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.671989] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.672106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.672118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.672255] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.097 [2024-07-15 18:12:48.672267] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.097 qpair failed and we were unable to recover it. 00:26:55.097 [2024-07-15 18:12:48.672426] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.672438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.672536] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.672548] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.672708] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.672720] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.672819] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.672830] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.672949] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.672961] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.673066] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.673077] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.673249] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.673261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.673453] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.673464] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.673648] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.673659] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.673820] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.673832] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.673981] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.673992] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.674100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.674111] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.674284] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.674297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.674404] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.674416] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.674595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.674607] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.674823] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.674836] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.674947] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.674960] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.675169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.675181] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.675346] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.675357] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.675488] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.675499] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.675754] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.675765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.675887] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.675899] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.676010] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.676021] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.676247] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.676269] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.676378] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.676389] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.676500] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.676511] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.676626] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.676637] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.676751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.676762] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.676932] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.676943] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.677055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.677067] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.677242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.677254] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.677368] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.677380] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.677567] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.677578] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.677741] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.677752] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.677849] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.677859] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.098 [2024-07-15 18:12:48.677961] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.098 [2024-07-15 18:12:48.677972] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.098 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.678144] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.678155] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.678213] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.678228] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.678342] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.678353] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.678527] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.678538] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.678704] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.678715] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.678826] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.678837] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.679001] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.679012] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.679134] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.679145] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.679247] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.679260] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.679356] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.679367] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.679484] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.679496] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.679673] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.679684] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.679781] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.679792] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.679961] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.679972] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.680075] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.680086] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.680250] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.680262] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.680418] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.680429] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.680626] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.680637] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.680799] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.680811] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.680926] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.680937] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.681053] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.681066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.681174] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.681186] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.681276] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.681289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.681390] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.681401] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.681566] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.681577] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.681664] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.681675] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.681800] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.681811] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.681898] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.681908] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.682113] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.682124] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.682283] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.682295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.682402] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.682414] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.682510] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.682522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.682670] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.682681] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.682788] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.682799] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.682910] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.682922] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.683032] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.683043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.683268] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.683280] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.683393] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.099 [2024-07-15 18:12:48.683405] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.099 qpair failed and we were unable to recover it. 00:26:55.099 [2024-07-15 18:12:48.683514] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.683525] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.683641] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.683652] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.683830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.683841] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.683928] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.683940] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.684099] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.684110] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.684229] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.684241] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.684411] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.684423] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.684531] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.684542] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.684717] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.684729] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.684842] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.684855] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.685000] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.685011] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.685173] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.685185] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.685344] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.685356] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.685461] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.685472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.685581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.685592] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.685692] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.685704] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.685815] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.685826] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.686057] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.686068] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.686207] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.686218] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.686332] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.686344] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.686503] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.686515] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.686707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.686718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.686946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.686960] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.687054] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.687066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.687162] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.687174] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.687283] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.687294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.687474] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.687486] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.687584] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.687595] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.687772] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.687783] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.687963] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.687974] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.688085] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.688097] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.688266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.688278] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.688399] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.688410] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.688521] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.688532] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.688643] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.688654] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.688765] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.688776] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.688887] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.688899] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.689077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.689089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.689278] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.689290] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.689369] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.689381] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.689488] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.689499] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.689622] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.689633] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.689729] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.689740] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.689842] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.689854] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.689956] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.689967] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.690134] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.690146] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.690241] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.690252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.690367] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.100 [2024-07-15 18:12:48.690378] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.100 qpair failed and we were unable to recover it. 00:26:55.100 [2024-07-15 18:12:48.690476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.690487] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.690582] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.690594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.690818] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.690829] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.690928] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.690940] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.691046] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.691057] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.691172] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.691183] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.691382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.691394] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.691492] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.691503] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.691671] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.691682] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.691852] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.691863] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.691964] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.691975] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.692075] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.692087] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.692249] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.692260] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.692441] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.692452] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.692522] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.692534] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.692713] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.692724] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.692834] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.692845] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.692969] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.692981] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.693235] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.693247] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.693442] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.693453] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.693588] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.693599] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.693721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.693732] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.693845] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.693856] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.694102] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.694113] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.694283] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.694294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.694452] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.694463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.694575] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.694587] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.694705] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.694716] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.694897] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.694909] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.695028] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.695041] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.695219] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.695245] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.695354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.695365] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.695523] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.695535] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.695715] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.695727] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.695826] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.695838] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.696028] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.696040] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.696192] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.696203] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.696287] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.696300] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.696526] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.696537] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.696635] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.696647] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.696844] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.696855] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.697083] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.697094] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.697321] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.697333] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.697466] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.697478] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.101 qpair failed and we were unable to recover it. 00:26:55.101 [2024-07-15 18:12:48.697637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.101 [2024-07-15 18:12:48.697648] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.697761] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.697773] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.697972] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.697984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.698089] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.698102] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.698213] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.698230] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.698402] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.698424] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.698585] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.698596] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.698700] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.698712] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.698841] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.698853] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.698969] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.698980] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.699206] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.699220] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.699331] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.699343] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.699518] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.699530] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.699704] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.699715] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.699834] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.699845] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.699952] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.699964] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.700062] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.700073] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.700251] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.700264] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.700432] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.700443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.700669] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.700680] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.700842] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.700853] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.701017] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.701029] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.701216] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.701231] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.701391] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.701403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.701571] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.701583] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.701697] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.701708] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.701884] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.701895] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.701994] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.702005] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.702240] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.702252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.702335] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.702346] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.702513] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.702524] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.702638] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.702649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.702757] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.702768] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.702999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.703010] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.703171] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.703182] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.703410] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.703422] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.703516] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.703528] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.703781] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.703793] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.703969] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.703980] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.704079] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.704090] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.704274] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.704285] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.704397] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.704409] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.704595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.704606] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.704801] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.704813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.704978] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.704990] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.705093] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.705104] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.705214] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.705234] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.705357] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.705369] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.705547] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.705559] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.705739] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.705750] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.705992] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.706005] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.102 qpair failed and we were unable to recover it. 00:26:55.102 [2024-07-15 18:12:48.706118] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.102 [2024-07-15 18:12:48.706130] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.706242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.706255] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.706352] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.706363] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.706588] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.706599] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.706766] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.706778] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.706997] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.707009] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.707111] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.707123] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.707236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.707247] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.707438] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.707450] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.707567] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.707578] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.707673] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.707685] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.707798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.707809] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.707914] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.707925] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.708092] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.708104] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.708209] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.708219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.708338] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.708351] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.708561] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.708573] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.708677] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.708689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.708880] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.708892] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.709014] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.709025] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.709216] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.709233] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.709497] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.709518] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.709601] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.709612] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.709822] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.709833] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.709953] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.709965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.710141] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.710153] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.710244] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.710259] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.710358] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.710370] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.710543] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.710555] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.710693] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.710704] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.710887] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.710899] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.711038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.711050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.711230] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.711242] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.711347] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.711359] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.711426] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.711437] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.711569] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.711581] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.711758] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.711769] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.711881] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.711893] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.711985] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.711997] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.712091] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.712102] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.712214] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.103 [2024-07-15 18:12:48.712230] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.103 qpair failed and we were unable to recover it. 00:26:55.103 [2024-07-15 18:12:48.712334] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.712345] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.712624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.712636] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.712810] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.712822] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.712940] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.712951] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.713248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.713261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.713453] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.713465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.713577] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.713589] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.713842] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.713854] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.713974] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.713986] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.714215] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.714231] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.714300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.714311] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.714396] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.714418] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.714587] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.714599] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.714785] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.714796] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.714969] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.714981] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.715173] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.715185] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.715379] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.715391] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.715492] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.715504] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.715614] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.715626] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.715790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.715801] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.715905] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.715916] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.716004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.716014] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.716178] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.716189] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.716377] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.716389] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.716550] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.716561] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.716650] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.716663] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.716769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.716780] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.716887] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.716899] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.716965] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.716975] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.717081] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.717093] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.717254] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.717266] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.717435] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.717446] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.717552] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.717564] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.717735] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.717746] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.717884] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.717896] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.718025] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.718038] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.718139] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.718151] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.718312] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.718324] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.718426] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.718438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.718545] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.718557] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.718733] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.104 [2024-07-15 18:12:48.718744] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.104 qpair failed and we were unable to recover it. 00:26:55.104 [2024-07-15 18:12:48.718857] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.718869] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.719037] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.719049] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.719166] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.719177] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.719334] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.719346] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.719511] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.719523] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.719637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.719649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.719754] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.719766] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.719879] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.719892] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.720055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.720066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.720192] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.720204] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.720308] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.720320] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.720464] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.720476] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.720652] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.720664] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.720841] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.720853] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.721031] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.721043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.721198] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.721209] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.721322] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.721335] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.721456] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.721467] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.721628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.721640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.721761] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.721773] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.721878] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.721891] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.721998] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.722010] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.722170] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.722183] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.722373] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.722385] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.722971] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.722997] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.723104] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.723116] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.723208] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.723219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.723391] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.723403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.723655] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.723668] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.723800] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.723812] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.723927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.723939] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.724039] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.724051] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.105 qpair failed and we were unable to recover it. 00:26:55.105 [2024-07-15 18:12:48.724192] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.105 [2024-07-15 18:12:48.724204] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.724380] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.724393] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.724536] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.724547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.724663] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.724675] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.724931] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.724943] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.725045] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.725057] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.725233] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.725246] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.725361] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.725373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.725500] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.725512] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.725616] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.725627] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.725767] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.725779] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.725876] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.725889] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.726081] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.726094] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.726266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.726279] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.726432] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.726444] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.726675] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.726687] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.726793] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.726805] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.726903] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.726915] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.727102] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.727114] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.727361] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.727373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.727469] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.727480] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.727708] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.727720] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.727893] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.727906] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.728142] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.728153] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.728312] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.728325] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.728577] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.728589] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.728692] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.728704] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.728814] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.728826] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.728947] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.728958] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.729058] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.729071] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.729286] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.729298] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.729384] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.729396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.729628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.729642] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.729754] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.729766] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.729872] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.729884] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.730109] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.730121] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.730295] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.730308] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.730391] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.730402] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.730505] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.730517] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.730713] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.730724] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.730835] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.730847] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.730952] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.730964] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.731137] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.731150] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.731259] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.731270] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.731467] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.731478] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.731583] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.731594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.731710] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.731722] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.731896] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.731908] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.732080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.106 [2024-07-15 18:12:48.732092] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.106 qpair failed and we were unable to recover it. 00:26:55.106 [2024-07-15 18:12:48.732201] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.732214] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.732328] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.732340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.732479] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.732491] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.732654] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.732665] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.732773] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.732784] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.732956] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.732967] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.733109] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.733120] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.733299] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.733311] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.733480] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.733492] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.733674] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.733686] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.733871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.733882] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.733966] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.733976] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.734077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.734090] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.734197] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.734210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.734378] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.734390] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.734556] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.734568] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.734732] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.734743] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.734858] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.734871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.734987] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.734999] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.735172] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.735184] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.735283] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.735295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.735386] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.735396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.735498] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.735514] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.735719] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.735734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.735893] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.735905] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.735991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.736002] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.736108] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.736120] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.736294] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.736308] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.736444] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.736455] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.736618] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.736628] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.736817] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.736829] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.736930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.736941] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.737055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.737066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.737166] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.737178] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.737285] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.737296] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.737461] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.737472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.737588] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.737601] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.737722] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.737735] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.737896] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.737908] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.738040] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.738052] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.738234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.738245] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.738347] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.738359] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.738567] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.738579] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.738681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.738693] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.738787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.738798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.739025] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.739038] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.739207] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.739219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.739340] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.739352] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.739591] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.739604] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.739852] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.739865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.740086] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.740099] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.740241] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.740254] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.740361] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.740373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.740500] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.740513] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.740612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.740624] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.740801] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.740813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.740938] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.740950] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.741069] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.741081] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.741261] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.741273] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.741387] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.741400] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.741553] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.741566] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.741679] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.741692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.741802] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.741815] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.741924] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.741938] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.742039] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.742051] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.742169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.107 [2024-07-15 18:12:48.742183] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.107 qpair failed and we were unable to recover it. 00:26:55.107 [2024-07-15 18:12:48.742361] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.742374] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.742550] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.742563] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.742731] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.742743] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.742925] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.742936] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.743124] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.743135] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.743326] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.743338] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.743515] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.743528] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.743694] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.743706] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.743807] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.743820] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.744005] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.744016] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.744146] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.744157] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.744268] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.744297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.744487] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.744500] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.744624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.744637] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.744735] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.744749] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.744848] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.744871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.744968] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.744979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.745083] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.745094] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.745283] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.745295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.745399] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.745411] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.745526] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.745537] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.745650] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.745663] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.745830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.745841] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.745945] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.745958] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.746046] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.746056] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.746271] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.746283] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.746466] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.746478] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.746594] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.746607] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.746681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.746692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.746800] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.746811] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.746980] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.746991] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.747156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.747168] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.747280] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.747293] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.747528] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.747540] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.747667] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.747681] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.747795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.747807] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.747920] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.747932] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.748029] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.748043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.748142] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.748155] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.748394] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.748407] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.748638] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.748649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.748753] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.748765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.748944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.748956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.749125] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.749137] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.749319] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.749331] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.749441] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.749453] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.749545] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.749556] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.749723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.749753] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.750026] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.750038] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.750152] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.750164] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.750325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.750338] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.750505] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.750517] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.750688] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.750700] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.750794] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.750804] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.750900] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.108 [2024-07-15 18:12:48.750911] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.108 qpair failed and we were unable to recover it. 00:26:55.108 [2024-07-15 18:12:48.751006] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.751018] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.751132] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.751144] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.751260] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.751289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.751468] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.751481] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.751638] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.751650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.751757] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.751770] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.751888] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.751911] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.752013] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.752025] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.752189] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.752201] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.752403] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.752440] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.752572] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.752589] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.752775] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.752790] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.752921] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.752937] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.753119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.753134] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.753246] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.753261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.753377] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.753392] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.753573] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.753588] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.753760] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.753775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.753887] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.753902] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.754002] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.754018] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.754112] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.754125] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.754316] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.754337] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.754517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.754529] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.754654] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.754666] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.754774] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.754786] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.754894] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.754906] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.755072] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.755084] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.755197] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.755210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.755303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.755314] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.755427] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.755440] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.755558] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.755570] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.755741] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.755753] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.755921] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.755933] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.756075] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.756087] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.756279] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.756291] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.756466] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.756478] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.756599] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.756611] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.756751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.756763] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.756869] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.756881] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.757055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.757066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.757233] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.757246] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.757477] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.757489] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.757667] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.757679] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.757783] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.757795] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.757920] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.757931] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.758111] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.758123] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.758358] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.758370] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.758495] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.758507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.758664] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.758676] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.758849] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.758864] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.758977] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.758989] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.759152] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.759164] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.759292] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.759304] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.759504] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.759515] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.759701] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.759712] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.759819] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.759831] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.760014] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.760026] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.760131] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.760143] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.760373] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.760396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.760511] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.760522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.760629] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.760640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.760765] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.760777] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.760953] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.760965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.761134] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.761146] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.761300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.761312] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.761551] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.109 [2024-07-15 18:12:48.761563] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.109 qpair failed and we were unable to recover it. 00:26:55.109 [2024-07-15 18:12:48.761668] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.110 [2024-07-15 18:12:48.761678] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.110 qpair failed and we were unable to recover it. 00:26:55.110 [2024-07-15 18:12:48.761886] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.110 [2024-07-15 18:12:48.761897] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.110 qpair failed and we were unable to recover it. 00:26:55.110 [2024-07-15 18:12:48.762031] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.110 [2024-07-15 18:12:48.762043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.110 qpair failed and we were unable to recover it. 00:26:55.110 [2024-07-15 18:12:48.762239] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.110 [2024-07-15 18:12:48.762251] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.110 qpair failed and we were unable to recover it. 00:26:55.110 [2024-07-15 18:12:48.762431] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.110 [2024-07-15 18:12:48.762442] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.110 qpair failed and we were unable to recover it. 00:26:55.110 [2024-07-15 18:12:48.762560] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.110 [2024-07-15 18:12:48.762572] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.110 qpair failed and we were unable to recover it. 00:26:55.383 [2024-07-15 18:12:48.762669] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.383 [2024-07-15 18:12:48.762683] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.383 qpair failed and we were unable to recover it. 00:26:55.383 [2024-07-15 18:12:48.762844] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.383 [2024-07-15 18:12:48.762857] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.383 qpair failed and we were unable to recover it. 00:26:55.383 [2024-07-15 18:12:48.763022] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.383 [2024-07-15 18:12:48.763034] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.383 qpair failed and we were unable to recover it. 00:26:55.383 [2024-07-15 18:12:48.763200] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.383 [2024-07-15 18:12:48.763212] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.383 qpair failed and we were unable to recover it. 00:26:55.383 [2024-07-15 18:12:48.763336] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.383 [2024-07-15 18:12:48.763347] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.383 qpair failed and we were unable to recover it. 00:26:55.383 [2024-07-15 18:12:48.763513] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.383 [2024-07-15 18:12:48.763525] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.383 qpair failed and we were unable to recover it. 00:26:55.383 [2024-07-15 18:12:48.763621] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.383 [2024-07-15 18:12:48.763633] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.383 qpair failed and we were unable to recover it. 00:26:55.383 [2024-07-15 18:12:48.763834] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.383 [2024-07-15 18:12:48.763845] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.383 qpair failed and we were unable to recover it. 00:26:55.383 [2024-07-15 18:12:48.763954] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.383 [2024-07-15 18:12:48.763967] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.383 qpair failed and we were unable to recover it. 00:26:55.383 [2024-07-15 18:12:48.764126] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.383 [2024-07-15 18:12:48.764138] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.383 qpair failed and we were unable to recover it. 00:26:55.383 [2024-07-15 18:12:48.764247] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.383 [2024-07-15 18:12:48.764259] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.383 qpair failed and we were unable to recover it. 00:26:55.383 [2024-07-15 18:12:48.764528] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.383 [2024-07-15 18:12:48.764539] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.383 qpair failed and we were unable to recover it. 00:26:55.383 [2024-07-15 18:12:48.764648] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.383 [2024-07-15 18:12:48.764659] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.383 qpair failed and we were unable to recover it. 00:26:55.383 [2024-07-15 18:12:48.764769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.383 [2024-07-15 18:12:48.764780] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.764917] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.764929] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.765037] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.765049] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.765149] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.765160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.765258] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.765272] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.765426] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.765438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.765607] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.765618] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.765709] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.765719] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.765901] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.765912] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.766016] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.766028] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.766247] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.766259] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.766375] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.766387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.766564] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.766576] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.766747] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.766758] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.766885] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.766896] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.766998] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.767010] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.767106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.767118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.767313] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.767325] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.767441] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.767454] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.767610] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.767622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.767699] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.767709] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.767816] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.767828] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.767944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.767956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.768124] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.768136] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.768248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.768261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.768358] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.768369] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.768473] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.768484] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.768556] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.768566] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.768660] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.768672] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.768766] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.768777] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.768954] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.768965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.769096] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.769107] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.769364] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.769377] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.769483] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.769496] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.769592] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.769604] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.769713] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.769725] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.769836] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.769847] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.769916] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.769927] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.769997] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.770007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.770178] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.770190] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.770275] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.770286] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.770378] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.770389] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.770484] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.770496] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.770611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.770623] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.770786] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.770799] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.770984] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.770996] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.771095] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.771107] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.771204] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.771215] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.771384] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.771396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.771576] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.771588] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.771720] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.771733] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.771828] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.771839] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.771960] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.771972] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.772076] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.772087] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.772190] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.772202] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.772314] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.772327] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.772500] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.772512] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.772598] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.772609] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.772842] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.384 [2024-07-15 18:12:48.772854] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.384 qpair failed and we were unable to recover it. 00:26:55.384 [2024-07-15 18:12:48.773014] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.773026] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.773200] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.773212] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.773377] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.773389] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.773500] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.773512] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.773685] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.773697] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.773957] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.773968] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.774070] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.774082] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.774191] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.774203] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.774303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.774315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.774502] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.774514] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.774625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.774637] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.774764] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.774775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.774883] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.774895] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.775124] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.775136] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.775311] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.775323] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.775437] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.775448] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.775657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.775668] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.775837] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.775848] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.776080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.776092] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.776261] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.776272] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.776433] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.776445] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.776544] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.776556] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.776719] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.776730] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.776838] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.776849] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.776951] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.776963] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.777129] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.777143] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.777266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.777278] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.777457] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.777468] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.777584] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.777596] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.777762] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.777773] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.777945] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.777957] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.778218] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.778234] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.778412] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.778425] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.778652] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.778663] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.778775] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.778787] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.778884] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.778895] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.778984] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.778995] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.779102] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.779114] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.779293] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.779306] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.779505] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.779517] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.779628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.779639] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.779795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.779807] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.779911] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.779922] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.779999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.780010] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.780122] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.780134] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.780257] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.780269] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.780430] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.780441] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.780552] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.780564] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.780676] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.780688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.780853] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.780865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.780957] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.780968] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.781130] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.781142] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.781375] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.781387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.781568] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.781579] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.781684] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.781696] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.781808] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.781820] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.385 [2024-07-15 18:12:48.781999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.385 [2024-07-15 18:12:48.782010] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.385 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.782079] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.782089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.782259] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.782271] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.782368] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.782379] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.782557] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.782569] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.782663] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.782674] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.782812] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.782823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.782999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.783011] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.783094] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.783106] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.783264] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.783279] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.783444] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.783456] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.783636] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.783649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.783762] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.783774] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.783874] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.783885] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.784049] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.784061] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.784168] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.784179] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.784352] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.784364] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.784467] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.784479] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.784598] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.784610] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.784717] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.784729] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.784954] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.784966] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.785143] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.785155] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.785279] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.785291] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.785459] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.785471] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.785588] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.785600] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.785712] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.785723] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.785811] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.785822] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.785996] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.786007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.786108] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.786120] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.786245] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.786258] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.786377] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.786389] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.786496] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.786508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.786671] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.786683] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.786900] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.786912] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.786984] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.786995] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.787177] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.787189] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.787262] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.787274] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.787391] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.787403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.787494] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.787506] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.787587] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.787599] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.787777] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.787788] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.787888] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.787901] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.788015] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.788026] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.788196] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.788208] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.788323] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.788335] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.788611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.788622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.788743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.788755] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.788922] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.788934] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.789105] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.789116] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.386 qpair failed and we were unable to recover it. 00:26:55.386 [2024-07-15 18:12:48.789288] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.386 [2024-07-15 18:12:48.789302] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.789512] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.789524] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.789624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.789635] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.789809] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.789820] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.789985] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.789997] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.790216] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.790232] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.790324] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.790334] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.790513] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.790524] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.790674] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.790686] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.790797] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.790809] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.791008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.791020] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.791135] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.791147] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.791307] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.791319] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.791477] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.791488] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.791604] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.791616] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.791778] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.791790] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.791965] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.791977] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.792138] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.792150] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.792262] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.792274] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.792377] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.792389] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.792506] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.792517] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.792613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.792624] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.792847] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.792860] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.792959] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.792971] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.793091] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.793103] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.793199] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.793210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.793324] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.793336] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.793456] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.793479] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.793599] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.793614] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.793746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.793762] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.793870] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.793885] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.794079] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.794094] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.794330] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.794348] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.794529] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.794545] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.794736] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.794752] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.794890] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.794905] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.795138] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.795153] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.795438] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.795454] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.795641] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.795658] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.795862] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.795877] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.796066] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.796081] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.796277] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.796292] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.796433] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.796449] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.796622] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.796637] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.796755] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.796769] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.797058] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.797074] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.797191] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.797207] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.797445] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.797461] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.797633] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.797648] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.797819] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.797835] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.797952] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.797967] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.798104] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.798119] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.798321] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.798336] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.798465] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.798480] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.798589] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.798607] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.798730] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.798746] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.798916] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.798931] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.799038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.799054] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.799162] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.799178] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.799381] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.799397] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.799507] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.799521] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.799763] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.387 [2024-07-15 18:12:48.799778] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.387 qpair failed and we were unable to recover it. 00:26:55.387 [2024-07-15 18:12:48.799951] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.799966] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.800145] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.800160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.800410] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.800426] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.800615] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.800631] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.800894] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.800908] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.801089] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.801104] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.801330] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.801346] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.801460] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.801476] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.801589] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.801604] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.801821] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.801836] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.801951] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.801966] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.802151] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.802166] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.802401] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.802416] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.802587] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.802602] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.802782] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.802798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.802905] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.802921] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.803035] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.803050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.803159] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.803174] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.803257] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.803271] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.803440] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.803458] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.803626] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.803641] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.803834] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.803849] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.804080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.804095] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.804331] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.804347] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.804522] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.804536] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.804716] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.804731] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.804984] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.804999] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.805168] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.805184] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.805300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.805316] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.805509] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.805524] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.805733] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.805748] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.805861] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.805877] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.806001] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.806016] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.806185] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.806201] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.806324] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.806340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.806524] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.806536] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.806638] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.806650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.806756] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.806767] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.806879] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.806890] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.807136] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.807148] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.807312] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.807324] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.807427] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.807439] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.807540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.807552] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.807641] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.807652] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.807744] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.807755] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.807870] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.807882] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.808042] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.808057] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.808230] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.808242] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.808419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.808430] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.808619] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.808631] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.808740] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.808751] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.808856] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.808867] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.809024] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.809036] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.809148] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.809160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.809263] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.809275] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.809380] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.809392] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.809561] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.809572] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.809752] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.809763] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.809923] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.809935] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.810107] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.810119] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.388 [2024-07-15 18:12:48.810299] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.388 [2024-07-15 18:12:48.810311] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.388 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.810495] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.810508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.810687] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.810699] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.810779] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.810789] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.810967] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.810979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.811093] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.811105] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.811367] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.811379] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.811492] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.811503] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.811621] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.811633] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.811798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.811809] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.811972] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.811983] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.812085] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.812096] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.812312] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.812325] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.812450] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.812462] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.812622] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.812634] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.812795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.812807] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.812915] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.812928] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.813051] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.813063] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.813196] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.813208] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.813374] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.813387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.813575] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.813587] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.813750] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.813761] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.813945] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.813957] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.814056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.814067] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.814174] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.814186] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.814344] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.814356] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.814469] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.814483] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.814579] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.814591] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.814706] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.814717] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.814813] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.814823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.814999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.815011] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.815184] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.815195] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.815300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.815313] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.815495] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.815507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.815628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.815640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.815906] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.815917] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.816098] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.816110] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.816234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.816246] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.816415] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.816427] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.816515] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.816526] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.816690] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.816702] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.816886] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.816898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.817151] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.817162] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.817333] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.817345] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.817507] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.817518] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.817611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.817623] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.817783] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.817795] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.818023] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.818036] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.818160] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.818172] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.818278] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.818290] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.818455] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.818466] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.818566] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.818577] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.818827] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.818839] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.818945] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.818957] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.819067] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.819079] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.389 [2024-07-15 18:12:48.819191] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.389 [2024-07-15 18:12:48.819203] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.389 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.819320] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.819333] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.819432] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.819443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.819552] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.819564] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.819714] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.819726] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.819869] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.819881] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.819980] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.819991] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.820074] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.820085] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.820252] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.820265] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.820360] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.820372] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.820577] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.820589] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.820749] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.820763] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.820857] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.820868] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.821039] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.821050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.821165] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.821176] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.821347] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.821360] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.821630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.821641] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.821801] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.821813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.822002] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.822014] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.822116] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.822127] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.822288] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.822300] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.822424] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.822435] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.822629] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.822641] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.822802] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.822814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.822917] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.822929] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.823119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.823130] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.823286] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.823299] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.823497] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.823508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.823618] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.823630] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.823813] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.823825] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.823991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.824002] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.824117] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.824129] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.824223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.824238] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.824380] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.824392] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.824492] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.824504] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.824606] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.824617] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.824735] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.824747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.824852] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.824864] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.825014] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.825026] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.825121] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.825132] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.825243] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.825254] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.825336] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.825346] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.825625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.825642] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.825810] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.825826] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.825959] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.825975] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.826090] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.826105] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.826238] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.826254] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.826372] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.826388] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.826575] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.826590] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.826746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.826761] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.826863] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.826878] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.827052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.827070] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.827269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.827286] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.827468] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.827483] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.827663] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.827679] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.827795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.827810] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.827950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.827966] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.828120] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.828136] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.828371] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.828386] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.828549] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.828564] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.390 [2024-07-15 18:12:48.828739] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.390 [2024-07-15 18:12:48.828754] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.390 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.828989] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.829004] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.829119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.829133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.829325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.829340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.829516] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.829531] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.829676] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.829710] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.829915] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.829928] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.830101] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.830113] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.830309] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.830322] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.830451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.830463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.830647] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.830659] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.830765] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.830777] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.830891] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.830903] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.831066] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.831078] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.831200] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.831212] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.831319] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.831331] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.831393] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.831403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.831572] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.831584] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.831686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.831700] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.831860] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.831871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.832045] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.832057] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.832219] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.832246] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.832343] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.832355] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.832556] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.832568] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.832676] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.832688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.832879] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.832891] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.833062] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.833074] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.833184] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.833196] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.833375] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.833387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.833485] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.833497] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.833597] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.833609] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.833782] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.833793] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.833980] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.833992] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.834116] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.834144] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.834268] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.834279] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.834386] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.834398] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.834511] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.834521] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.834746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.834757] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.834853] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.834865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.835111] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.835122] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.835234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.835247] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.835454] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.835465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.835562] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.835572] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.835822] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.835834] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.835947] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.835959] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.836154] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.836173] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.836292] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.836308] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.836428] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.836443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.836616] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.836631] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.836745] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.836760] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.836872] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.836887] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.837001] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.837016] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.837200] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.837215] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.837321] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.837337] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.837566] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.837581] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.837743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.837758] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.837928] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.837943] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.838070] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.838086] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.838274] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.838293] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.838466] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.838481] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.838770] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.838784] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.838971] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.838986] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.839103] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.839119] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.391 qpair failed and we were unable to recover it. 00:26:55.391 [2024-07-15 18:12:48.839322] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.391 [2024-07-15 18:12:48.839337] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.839507] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.839522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.839631] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.839646] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.839814] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.839828] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.839951] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.839967] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.840084] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.840098] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.840216] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.840236] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.840352] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.840366] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.840547] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.840563] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.840689] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.840705] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.840813] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.840828] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.840936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.840950] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.841055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.841070] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.841203] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.841218] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.841399] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.841414] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.841537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.841553] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.841722] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.841738] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.841973] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.841988] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.842189] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.842203] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.842388] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.842404] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.842581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.842596] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.842769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.842784] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.842997] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.843031] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.843219] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.843246] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.843378] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.843391] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.843502] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.843514] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.843686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.843698] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.843789] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.843800] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.843907] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.843918] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.844165] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.844177] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.844421] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.844433] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.844553] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.844566] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.844751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.844763] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.844875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.844887] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.845069] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.845081] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.845242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.845257] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.845375] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.845387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.845595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.845606] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.845712] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.845723] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.845903] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.845915] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.846012] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.846023] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.846144] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.846155] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.846335] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.846346] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.846584] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.846595] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.846777] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.846788] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.846910] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.846921] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.847096] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.847108] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.847206] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.847217] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.847400] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.847413] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.847657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.847669] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.847772] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.847784] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.847902] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.847914] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.848014] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.848025] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.848222] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.848239] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.392 [2024-07-15 18:12:48.848343] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.392 [2024-07-15 18:12:48.848355] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.392 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.848518] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.848529] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.848709] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.848721] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.848886] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.848897] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.849026] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.849037] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.849215] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.849234] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.849420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.849432] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.849544] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.849572] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.849687] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.849708] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.849830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.849847] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.850036] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.850054] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.850242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.850256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.850486] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.850497] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.850594] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.850605] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.850804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.850815] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.850931] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.850942] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.851135] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.851147] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.851328] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.851340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.851498] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.851511] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.851686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.851698] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.851823] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.851834] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.851940] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.851954] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.852065] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.852077] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.852187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.852199] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.852391] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.852414] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.852598] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.852610] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.852816] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.852827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.852993] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.853005] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.853117] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.853129] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.853326] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.853338] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.853426] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.853436] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.853565] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.853577] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.853748] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.853760] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.853931] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.853942] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.854052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.854063] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.854172] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.854184] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.854279] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.854289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.854377] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.854388] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.854477] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.854488] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.854591] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.854602] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.854795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.854807] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.855031] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.855042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.855211] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.855223] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.855398] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.855410] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.855510] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.855522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.855630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.855641] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.855753] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.855765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.855881] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.855892] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.856069] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.856087] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.856263] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.856279] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.856448] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.856463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.856592] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.856608] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.856812] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.856827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.856931] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.856946] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.857129] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.857141] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.857321] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.857333] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.857439] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.857451] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.393 [2024-07-15 18:12:48.857562] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.393 [2024-07-15 18:12:48.857573] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.393 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.857743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.857754] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.857830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.857840] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.857943] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.857954] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.858140] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.858151] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.858344] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.858356] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.858527] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.858538] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.858650] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.858681] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.858971] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.859002] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.859217] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.859256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.859470] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.859501] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.859652] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.859683] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.859839] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.859870] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.860080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.860110] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.860379] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.860410] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.860552] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.860564] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.860776] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.860806] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.861028] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.861058] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.861293] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.861326] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.861466] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.861477] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.861653] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.861684] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.861901] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.861931] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.862132] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.862162] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.862436] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.862467] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.862662] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.862673] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.862927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.862958] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.863076] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.863106] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.863309] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.863340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.863562] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.863592] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.863801] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.863832] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.864041] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.864072] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.864263] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.864300] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.864517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.864548] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.864749] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.864780] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.864937] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.864967] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.865172] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.865203] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.865359] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.865390] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.865682] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.865713] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.865923] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.865934] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.866130] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.866160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.866433] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.866465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.866735] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.866765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.866928] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.866959] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.867175] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.867206] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.867359] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.867390] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.867536] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.867567] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.867810] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.867822] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.868035] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.868065] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.868277] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.868308] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.868518] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.868549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.868795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.868825] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.869030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.869042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.869142] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.869154] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.869332] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.869344] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.869580] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.869611] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.869747] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.869778] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.870004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.870034] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.870252] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.870283] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.870490] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.870502] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.870681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.870693] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.870830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.870862] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.394 [2024-07-15 18:12:48.871003] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.394 [2024-07-15 18:12:48.871033] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.394 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.871168] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.871199] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.871428] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.871459] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.871629] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.871660] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.871878] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.871909] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.872135] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.872165] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.872408] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.872440] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.872671] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.872702] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.872901] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.872912] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.873082] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.873113] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.873387] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.873427] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.873623] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.873635] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.873770] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.873800] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.873948] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.873979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.874183] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.874214] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.874497] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.874528] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.874737] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.874767] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.874970] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.874981] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.875186] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.875197] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.875373] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.875385] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.875552] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.875564] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.875689] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.875719] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.875930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.875960] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.876170] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.876200] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.876435] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.876467] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.876741] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.876752] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.876908] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.876939] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.877137] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.877168] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.877351] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.877382] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.877528] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.877559] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.877701] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.877731] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.877885] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.877897] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.878079] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.878109] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.878313] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.878344] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.878474] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.878505] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.878705] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.878717] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.878841] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.878852] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.879091] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.879103] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.879211] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.879223] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.879297] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.879307] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.879450] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.879461] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.879587] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.879616] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.879768] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.879798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.880116] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.880148] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.880364] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.880397] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.880633] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.880663] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.880855] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.880867] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.881057] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.881087] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.881302] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.881334] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.881559] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.881570] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.881736] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.881771] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.881928] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.881958] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.882169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.882200] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.882342] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.882373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.882620] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.882651] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.882870] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.395 [2024-07-15 18:12:48.882902] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.395 qpair failed and we were unable to recover it. 00:26:55.395 [2024-07-15 18:12:48.883053] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.883084] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.883218] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.883256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.883471] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.883501] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.883645] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.883676] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.883942] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.883973] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.884264] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.884296] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.884467] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.884498] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.884662] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.884693] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.884907] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.884938] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.885205] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.885243] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.885396] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.885427] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.885573] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.885611] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.885839] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.885851] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.885946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.885957] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.886086] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.886098] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.886247] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.886259] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.886456] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.886486] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.886630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.886661] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.886878] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.886912] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.887109] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.887121] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.887240] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.887251] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.887460] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.887491] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.887761] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.887792] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.888017] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.888047] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.888257] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.888289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.888440] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.888471] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.888624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.888654] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.888912] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.888923] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.889103] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.889125] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.889265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.889278] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.889388] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.889400] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.889570] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.889600] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.889793] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.889823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.890029] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.890059] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.890355] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.890396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.890512] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.890524] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.890697] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.890727] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.890964] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.890995] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.891137] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.891167] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.891459] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.891491] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.891651] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.891678] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.891915] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.891946] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.892105] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.892135] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.892284] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.892317] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.892591] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.892621] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.892771] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.892802] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.893004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.893034] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.893184] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.893215] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.893499] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.893530] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.893734] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.893765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.894035] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.894066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.894393] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.894424] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.894721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.894751] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.894903] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.894933] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.895233] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.895245] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.895365] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.895377] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.895557] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.895568] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.895754] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.895785] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.895941] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.895972] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.396 [2024-07-15 18:12:48.896093] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.396 [2024-07-15 18:12:48.896124] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.396 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.896339] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.896371] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.896542] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.896572] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.896824] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.896836] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.897007] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.897019] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.897244] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.897255] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.897483] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.897495] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.897599] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.897611] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.897783] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.897794] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.897959] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.897971] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.898097] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.898127] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.898370] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.898401] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.898561] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.898593] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.898794] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.898806] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.898972] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.898984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.899088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.899102] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.899188] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.899199] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.899376] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.899407] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.899603] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.899633] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.899861] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.899901] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.900012] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.900022] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.900244] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.900276] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.900441] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.900472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.900680] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.900711] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.900976] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.901007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.901169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.901200] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.901448] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.901480] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.901619] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.901650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.901799] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.901829] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.901976] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.902019] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.902093] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.902103] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.902296] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.902328] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.902536] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.902567] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.902836] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.902866] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.903114] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.903126] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.903303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.903315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.903518] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.903549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.903778] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.903790] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.904017] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.904029] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.904142] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.904154] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.904316] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.904328] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.904443] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.904473] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.904640] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.904671] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.904892] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.904922] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.905190] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.905221] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.905429] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.905441] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.905545] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.905556] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.905734] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.905764] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.906066] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.906097] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.906396] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.906428] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.906584] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.906615] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.906797] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.906808] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.906916] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.906944] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.907160] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.907190] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.907338] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.907369] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.907582] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.907618] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.907744] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.907775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.907957] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.907969] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.908078] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.908109] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.908276] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.908309] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.908470] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.908501] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.397 qpair failed and we were unable to recover it. 00:26:55.397 [2024-07-15 18:12:48.908701] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.397 [2024-07-15 18:12:48.908731] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.908935] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.908966] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.909267] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.909298] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.909586] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.909617] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.909833] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.909864] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.910157] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.910187] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.910420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.910451] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.910657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.910688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.910861] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.910892] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.911110] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.911140] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.911296] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.911328] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.911598] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.911628] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.911868] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.911879] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.911985] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.911995] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.912186] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.912217] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.912448] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.912480] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.912686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.912717] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.912968] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.912979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.913142] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.913153] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.913269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.913300] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.913461] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.913492] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.913786] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.913817] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.913953] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.913984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.914189] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.914220] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.914475] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.914506] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.914707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.914738] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.915006] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.915037] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.915245] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.915277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.915453] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.915483] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.915640] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.915651] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.915761] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.915774] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.915958] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.915969] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.916195] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.916207] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.916468] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.916499] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.916726] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.916762] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.916992] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.917022] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.917292] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.917323] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.917496] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.917526] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.917740] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.917770] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.917907] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.917938] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.918142] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.918172] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.918332] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.918364] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.918557] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.918569] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.918665] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.918676] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.918873] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.918904] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.919123] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.919154] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.919319] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.919350] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.919553] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.919584] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.919797] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.919827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.919974] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.920005] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.920304] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.920336] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.920496] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.398 [2024-07-15 18:12:48.920528] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.398 qpair failed and we were unable to recover it. 00:26:55.398 [2024-07-15 18:12:48.920750] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.920780] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.920983] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.921014] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.921218] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.921256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.921471] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.921502] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.921712] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.921742] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.921907] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.921937] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.922235] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.922267] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.922424] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.922455] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.922665] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.922676] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.922845] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.922876] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.923096] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.923127] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.923295] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.923326] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.923614] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.923644] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.923858] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.923888] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.924106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.924135] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.924357] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.924389] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.924657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.924688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.924902] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.924933] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.925128] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.925140] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.925306] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.925318] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.925514] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.925545] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.925812] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.925842] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.926062] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.926097] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.926400] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.926432] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.926729] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.926759] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.926905] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.926936] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.927157] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.927189] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.927467] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.927498] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.927681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.927692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.927795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.927805] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.928064] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.928095] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.928338] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.928384] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.928596] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.928626] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.928772] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.928802] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.929053] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.929064] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.929241] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.929252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.929511] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.929542] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.929779] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.929809] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.930075] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.930087] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.930205] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.930217] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.930396] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.930428] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.930634] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.930664] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.930823] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.930853] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.931072] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.931102] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.931330] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.931362] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.931584] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.931614] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.931829] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.931858] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.932023] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.932053] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.932345] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.932377] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.932592] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.932623] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.932763] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.932793] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.933008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.933038] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.933191] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.933221] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.933472] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.933503] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.933740] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.933771] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.933907] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.933938] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.934123] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.934135] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.399 [2024-07-15 18:12:48.934363] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.399 [2024-07-15 18:12:48.934375] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.399 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.934557] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.934587] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.934742] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.934773] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.934971] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.935001] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.935209] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.935247] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.935387] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.935426] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.935719] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.935751] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.936058] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.936088] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.936301] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.936333] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.936563] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.936593] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.936812] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.936825] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.937100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.937130] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.937282] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.937313] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.937628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.937659] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.937870] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.937882] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.938056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.938068] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.938282] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.938304] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.938468] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.938479] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.938608] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.938639] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.938848] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.938879] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.939097] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.939127] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.939335] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.939366] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.939531] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.939562] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.939709] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.939740] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.939918] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.939948] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.940155] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.940166] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.940305] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.940319] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.940497] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.940528] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.940751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.940782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.941003] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.941034] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.941198] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.941236] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.941459] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.941489] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.941764] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.941801] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.941986] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.942003] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.942187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.942203] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.942334] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.942350] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.942609] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.942625] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.942796] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.942811] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.943032] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.943063] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.943353] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.943384] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.943677] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.943709] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.943910] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.943942] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.944157] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.944172] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.944354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.944370] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.944488] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.944504] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.944639] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.944659] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.944831] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.944847] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.944969] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.944984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.945181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.945212] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.945388] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.945420] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.945630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.945662] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.945807] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.945822] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.946069] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.946103] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.946260] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.946292] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.946507] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.946539] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.946680] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.946691] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.946865] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.946896] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.947193] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.947246] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.947464] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.947497] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.947776] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.947806] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.400 [2024-07-15 18:12:48.948024] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.400 [2024-07-15 18:12:48.948054] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.400 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.948323] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.948354] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.948506] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.948537] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.948756] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.948788] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.948991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.949022] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.949276] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.949289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.949408] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.949421] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.949582] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.949594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.949854] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.949884] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.950045] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.950076] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.950242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.950273] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.950488] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.950518] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.950782] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.950819] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.951016] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.951034] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.951760] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.951786] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.951986] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.952018] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.952311] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.952345] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.952982] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.953003] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.953194] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.953211] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.953478] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.953493] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.953683] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.953699] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.953905] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.953920] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.954102] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.954118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.954197] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.954212] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.954318] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.954331] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.954562] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.954577] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.954675] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.954685] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.954858] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.954871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.954966] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.954976] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.955170] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.955182] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.955284] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.955295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.955474] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.955485] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.955661] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.955672] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.955839] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.955851] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.955959] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.955971] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.956134] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.956145] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.956358] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.956390] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.956609] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.956640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.956850] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.956880] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.957011] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.957023] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.957132] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.957143] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.957372] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.957403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.957554] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.957585] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.957737] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.957780] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.957961] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.957973] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.958194] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.958233] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.958441] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.958472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.958677] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.958707] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.958911] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.958941] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.959187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.959218] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.959369] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.959400] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.959703] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.959734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.959949] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.959961] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.960068] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.401 [2024-07-15 18:12:48.960079] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.401 qpair failed and we were unable to recover it. 00:26:55.401 [2024-07-15 18:12:48.960241] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.960254] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.960376] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.960387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.960552] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.960564] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.960791] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.960823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.960990] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.961020] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.961222] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.961261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.961419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.961450] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.961681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.961712] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.961974] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.961986] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.962286] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.962319] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.962544] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.962574] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.962788] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.962823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.963030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.963043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.963147] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.963176] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.963348] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.963380] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.963534] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.963564] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.963850] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.963863] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.964068] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.964099] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.964374] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.964406] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.964548] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.964579] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.964717] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.964747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.964888] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.964918] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.965130] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.965161] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.965325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.965356] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.965514] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.965545] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.965700] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.965732] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.965951] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.965981] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.966193] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.966232] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.966530] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.966561] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.966716] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.966747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.966923] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.966954] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.967166] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.967178] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.967410] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.967422] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.967538] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.967568] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.967773] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.967803] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.968047] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.968078] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.968292] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.968324] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.968452] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.968483] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.968752] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.968823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.968988] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.969022] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.969273] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.969307] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.969610] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.969642] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.969865] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.969896] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.970117] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.970148] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.970305] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.970337] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.970481] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.970511] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.970675] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.970706] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.970975] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.971006] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.971119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.971150] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.971373] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.971405] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.971609] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.971641] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.971854] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.971885] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.972050] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.972081] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.972288] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.972321] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.972484] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.972515] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.972741] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.972772] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.972999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.973014] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.973197] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.973213] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.402 qpair failed and we were unable to recover it. 00:26:55.402 [2024-07-15 18:12:48.973400] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.402 [2024-07-15 18:12:48.973416] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.973596] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.973627] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.973854] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.973885] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.974106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.974137] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.974380] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.974412] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.974642] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.974674] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.974859] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.974889] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.975193] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.975241] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.975457] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.975488] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.975694] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.975725] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.975893] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.975924] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.976146] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.976177] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.976422] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.976455] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.976678] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.976709] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.976936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.976967] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.977265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.977297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.977445] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.977476] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.977683] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.977715] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.977867] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.977898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.978100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.978132] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.978451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.978482] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.978774] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.978805] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.979010] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.979026] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.979197] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.979242] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.979544] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.979575] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.979746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.979761] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.979898] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.979928] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.980082] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.980113] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.980411] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.980443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.980675] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.980706] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.980879] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.980910] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.981103] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.981118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.981289] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.981305] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.981492] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.981523] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.981763] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.981800] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.982067] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.982098] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.982306] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.982337] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.982605] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.982636] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.982790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.982822] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.983044] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.983075] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.983243] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.983259] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.983357] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.983394] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.983614] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.983644] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.983823] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.983854] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.984162] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.984178] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.984376] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.984392] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.984583] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.984598] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.984830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.984846] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.984970] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.984986] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.985181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.985212] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.985442] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.985474] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.985620] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.985651] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.985872] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.985887] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.986054] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.986069] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.986201] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.986216] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.986346] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.986361] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.986537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.986552] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.986731] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.986747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.986950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.986982] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.987188] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.987219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.987519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.403 [2024-07-15 18:12:48.987549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.403 qpair failed and we were unable to recover it. 00:26:55.403 [2024-07-15 18:12:48.987707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.987738] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.987948] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.987979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.988160] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.988176] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.988376] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.988408] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.988632] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.988664] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.988856] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.988888] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.989197] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.989213] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.989432] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.989448] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.989702] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.989733] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.989940] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.989971] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.990262] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.990294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.990567] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.990597] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.990913] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.990945] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.991222] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.991282] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.991508] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.991540] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.991878] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.991894] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.992078] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.992094] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.992219] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.992241] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.992401] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.992417] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.992652] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.992667] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.992854] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.992870] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.993152] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.993167] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.993421] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.993437] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.993624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.993640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.993909] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.993924] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.994161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.994176] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.994355] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.994371] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.994634] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.994665] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.994875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.994906] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.995110] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.995140] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.995419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.995451] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.995704] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.995735] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.995944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.995976] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.996269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.996301] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.996531] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.996562] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.996804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.996836] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.997029] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.997045] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.997219] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.997262] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.997502] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.997533] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.997744] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.997775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.997989] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.998021] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.998247] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.998284] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.998634] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.998666] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.998876] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.998907] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.999252] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.999285] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.999523] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.999555] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:48.999767] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:48.999798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:49.000065] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:49.000097] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:49.000376] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:49.000408] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:49.000646] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:49.000677] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:49.000962] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:49.000978] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:49.001143] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:49.001159] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:49.001420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:49.001453] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:49.001627] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:49.001657] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.404 [2024-07-15 18:12:49.001940] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.404 [2024-07-15 18:12:49.001971] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.404 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.002273] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.002306] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.002553] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.002584] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.002869] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.002901] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.003127] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.003143] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.003328] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.003344] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.003584] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.003615] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.003904] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.003936] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.004263] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.004297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.004612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.004643] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.004883] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.004898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.005130] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.005145] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.005337] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.005353] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.005587] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.005603] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.005839] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.005857] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.006122] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.006162] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.006385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.006417] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.006697] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.006727] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.007042] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.007073] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.007355] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.007386] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.007670] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.007701] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.007908] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.007939] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.008144] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.008175] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.008461] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.008493] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.008763] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.008794] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.009147] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.009178] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.009431] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.009463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.009685] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.009717] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.009951] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.009982] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.010193] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.010235] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.010530] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.010562] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.010857] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.010888] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.011188] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.011220] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.011516] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.011548] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.011815] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.011848] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.012087] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.012119] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.012377] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.012394] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.012652] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.012667] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.012842] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.012859] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.013117] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.013147] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.013448] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.013481] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.013775] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.013811] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.014104] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.014136] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.014342] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.014374] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.014580] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.014611] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.014880] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.014919] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.015111] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.015127] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.015299] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.015315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.015522] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.015553] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.015857] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.015888] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.016107] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.016138] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.016407] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.016441] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.016646] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.016677] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.016897] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.016928] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.017219] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.017261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.017549] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.017581] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.017784] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.017815] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.018164] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.018196] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.018476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.018508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.018805] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.018837] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.019005] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.019037] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.405 [2024-07-15 18:12:49.019265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.405 [2024-07-15 18:12:49.019297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.405 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.019523] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.019555] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.019848] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.019879] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.020113] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.020144] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.020474] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.020507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.020748] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.020780] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.021117] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.021148] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.021383] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.021416] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.021715] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.021747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.022040] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.022079] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.022325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.022357] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.022674] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.022705] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.022985] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.023016] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.023252] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.023268] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.023454] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.023470] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.023759] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.023791] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.024102] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.024118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.024416] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.024449] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.024612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.024643] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.024844] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.024875] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.025184] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.025215] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.025504] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.025536] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.025835] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.025867] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.026159] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.026174] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.026474] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.026506] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.026825] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.026857] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.027126] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.027141] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.027347] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.027363] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.027543] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.027559] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.027833] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.027848] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.028029] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.028044] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.028220] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.028272] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.028497] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.028529] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.028717] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.028748] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.029019] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.029050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.029366] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.029382] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.029575] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.029591] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.029827] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.029844] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.030126] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.030142] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.030333] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.030350] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.030617] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.030648] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.030927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.030958] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.031271] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.031303] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.031521] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.031551] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.031829] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.031860] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.032098] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.032113] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.032363] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.032387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.032669] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.032701] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.033003] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.033046] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.033336] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.033369] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.033663] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.033694] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.033981] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.034012] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.034247] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.034280] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.034578] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.034609] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.034900] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.034931] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.035073] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.035103] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.035306] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.035338] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.035636] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.035667] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.035869] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.035901] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.036145] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.036176] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.036454] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.036470] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.406 [2024-07-15 18:12:49.036727] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.406 [2024-07-15 18:12:49.036767] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.406 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.037067] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.037098] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.037392] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.037425] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.037630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.037661] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.037977] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.038007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.038300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.038316] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.038525] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.038540] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.038728] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.038743] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.039002] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.039046] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.039336] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.039368] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.039687] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.039718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.040004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.040035] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.040263] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.040280] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.040468] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.040483] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.040774] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.040810] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.041091] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.041122] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.041369] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.041385] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.041580] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.041611] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.041882] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.041914] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.042081] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.042113] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.042335] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.042350] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.042535] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.042566] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.042865] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.042898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.043108] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.043125] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.043374] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.043390] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.043657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.043688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.043904] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.043920] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.044211] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.044251] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.044556] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.044587] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.044872] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.044903] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.045202] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.045243] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.045471] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.045502] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.045658] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.045689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.045991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.046022] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.046314] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.046346] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.046638] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.046669] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.046891] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.046923] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.047214] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.047261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.047443] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.047458] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.047722] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.047737] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.047990] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.048022] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.048292] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.048325] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.048633] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.048665] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.048980] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.049011] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.049223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.049265] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.049527] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.049544] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.049760] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.049776] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.049982] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.049998] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.050236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.050252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.050423] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.050439] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.050651] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.050682] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.050890] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.050921] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.051132] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.051163] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.051450] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.407 [2024-07-15 18:12:49.051466] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.407 qpair failed and we were unable to recover it. 00:26:55.407 [2024-07-15 18:12:49.051707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.051723] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.051912] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.051929] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.052200] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.052241] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.052494] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.052526] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.052845] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.052876] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.053115] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.053147] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.053445] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.053477] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.053713] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.053745] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.053895] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.053910] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.054178] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.054210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.054432] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.054464] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.054754] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.054785] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.055057] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.055089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.055378] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.055394] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.055572] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.055589] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.055795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.055827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.056123] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.056154] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.056454] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.056487] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.056787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.056818] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.057055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.057086] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.057380] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.057412] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.057684] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.057715] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.057916] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.057954] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.058243] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.058276] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.058581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.058613] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.058914] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.058946] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.059148] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.059180] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.059417] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.059450] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.059747] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.059788] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.060058] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.060074] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.060247] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.060263] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.060451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.060467] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.060723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.060739] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.060928] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.060944] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.061157] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.061174] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.061436] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.061451] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.061646] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.061662] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.061941] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.061973] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.062295] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.062327] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.062624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.062655] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.062811] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.062843] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.063138] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.063170] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.063469] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.063501] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.063799] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.063830] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.064128] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.064172] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.064386] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.064403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.064646] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.064663] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.064901] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.064917] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.065154] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.065170] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.065340] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.065357] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.065563] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.065595] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.065885] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.065916] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.066222] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.066243] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.066442] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.066457] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.066632] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.066648] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.066936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.066972] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.067250] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.067293] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.067494] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.067511] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.067790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.067822] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.068046] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.068077] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.068285] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.068319] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.068643] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.408 [2024-07-15 18:12:49.068674] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.408 qpair failed and we were unable to recover it. 00:26:55.408 [2024-07-15 18:12:49.068912] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.068944] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.069248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.069280] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.069591] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.069622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.069927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.069958] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.070207] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.070246] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.070404] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.070419] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.070686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.070717] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.071015] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.071046] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.071347] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.071379] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.071654] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.071685] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.071900] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.071931] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.072139] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.072170] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.072386] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.072403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.072592] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.072623] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.072921] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.072952] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.073121] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.073138] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.073378] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.073394] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.073665] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.073697] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.073988] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.074020] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.074267] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.074284] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.074395] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.074413] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.074679] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.074711] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.074928] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.074959] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.075173] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.075204] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.075425] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.075457] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.075680] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.075712] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.075984] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.076024] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.076201] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.076217] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.076418] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.076434] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.076709] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.076740] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.076903] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.076934] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.077074] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.077106] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.077332] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.077366] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.077638] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.077670] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.078039] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.078111] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.078389] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.078420] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.078693] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.078707] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.078993] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.079025] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.079351] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.079385] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.079688] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.079721] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.080041] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.080072] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.080288] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.080300] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.080490] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.080502] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.080741] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.080773] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.081073] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.081104] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.081398] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.081411] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.081596] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.081608] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.081844] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.081885] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.082105] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.082136] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.082375] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.082387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.082625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.082656] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.082871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.082903] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.083198] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.083240] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.083404] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.083436] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.083721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.083752] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.083970] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.084001] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.084250] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.084283] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.084522] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.084553] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.084836] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.409 [2024-07-15 18:12:49.084869] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.409 qpair failed and we were unable to recover it. 00:26:55.409 [2024-07-15 18:12:49.085167] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.085199] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.085522] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.085593] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.085851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.085893] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.086078] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.086094] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.086297] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.086329] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.086626] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.086657] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.086958] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.087002] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.087181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.087197] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.087492] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.087509] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.087754] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.087770] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.087941] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.087957] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.088153] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.088169] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.088362] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.088394] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.088630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.088663] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.088957] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.088988] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.089285] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.089323] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.089646] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.089676] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.089900] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.089932] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.090244] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.090277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.090559] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.090591] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.090866] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.090898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.091174] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.091189] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.091406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.091422] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.091601] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.091616] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.091902] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.091934] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.092222] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.092250] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.092550] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.092582] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.092796] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.092828] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.093123] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.093165] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.093416] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.093432] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.093661] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.093678] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.093956] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.093988] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.094251] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.094298] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.094610] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.094642] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.094853] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.094884] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.095176] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.095191] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.095322] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.095338] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.410 [2024-07-15 18:12:49.095483] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.410 [2024-07-15 18:12:49.095499] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.410 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.095774] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.095791] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.095979] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.095995] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.096256] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.096272] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.096530] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.096561] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.096838] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.096876] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.097171] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.097202] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.097499] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.097532] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.097718] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.097749] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.097970] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.098001] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.098286] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.098322] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.098517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.098549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.098709] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.098740] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.099049] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.099065] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.099240] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.099257] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.099501] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.099532] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.099691] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.099723] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.100065] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.100097] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.100323] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.100356] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.100688] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.100721] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.100945] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.100976] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.101281] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.101298] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.101591] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.101623] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.101948] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.101979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.102234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.102250] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.688 [2024-07-15 18:12:49.102491] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.688 [2024-07-15 18:12:49.102507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.688 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.102798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.102829] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.103146] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.103177] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.103494] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.103527] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.103749] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.103781] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.104026] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.104058] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.104254] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.104271] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.104559] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.104575] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.104821] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.104836] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.105008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.105025] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.105197] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.105212] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.105482] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.105499] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.105688] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.105705] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.105856] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.105888] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.106096] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.106112] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.106381] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.106413] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.106704] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.106736] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.106975] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.107007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.107302] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.107318] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.107568] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.107584] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.107879] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.107910] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.108126] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.108158] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.108439] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.108477] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.108729] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.108745] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.108985] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.109001] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.109261] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.109293] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.109593] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.109625] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.109788] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.109819] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.110077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.110109] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.110411] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.110443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.110740] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.110772] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.111072] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.111103] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.111276] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.111308] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.111517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.111549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.111825] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.111856] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.112160] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.112177] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.112382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.112399] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.112597] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.112629] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.112769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.112801] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.113034] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.113065] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.113364] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.113381] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.689 [2024-07-15 18:12:49.113606] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.689 [2024-07-15 18:12:49.113622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.689 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.113795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.113811] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.113992] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.114024] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.114316] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.114332] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.114612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.114644] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.114894] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.114925] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.115220] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.115261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.115507] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.115543] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.115877] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.115908] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.116062] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.116094] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.116378] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.116394] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.116640] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.116655] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.116842] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.116858] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.117141] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.117173] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.117490] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.117522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.117795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.117827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.118058] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.118089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.118309] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.118341] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.118545] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.118560] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.118753] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.118785] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.119002] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.119033] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.119315] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.119347] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.119622] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.119653] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.119969] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.120000] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.120269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.120303] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.120557] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.120589] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.120863] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.120894] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.121197] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.121238] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.121463] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.121495] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.121743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.121774] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.121918] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.121949] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.122115] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.122149] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.122431] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.122463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.122653] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.122669] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.122961] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.123003] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.123280] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.123311] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.123625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.123657] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.123944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.123975] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.124288] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.124321] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.124539] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.124570] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.124785] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.124816] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.125045] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.690 [2024-07-15 18:12:49.125077] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.690 qpair failed and we were unable to recover it. 00:26:55.690 [2024-07-15 18:12:49.125301] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.125334] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.125608] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.125639] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.125855] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.125885] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.126164] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.126195] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.126415] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.126447] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.126653] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.126684] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.127034] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.127066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.127312] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.127344] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.127524] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.127555] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.127801] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.127832] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.128113] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.128144] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.128351] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.128384] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.128621] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.128652] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.128877] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.128910] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.129207] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.129251] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.129485] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.129516] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.129797] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.129813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.129993] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.130009] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.130272] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.130305] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.130536] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.130572] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.130848] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.130879] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.131105] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.131136] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.131439] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.131472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.131699] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.131731] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.132034] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.132065] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.132311] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.132344] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.132503] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.132519] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.132818] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.132850] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.133116] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.133147] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.133473] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.133507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.133721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.133752] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.133897] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.133929] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.134199] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.134215] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.134396] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.134412] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.134577] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.134592] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.134733] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.134749] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.135012] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.135044] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.135221] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.135262] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.135538] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.135554] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.135775] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.135807] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.136106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.136137] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.691 qpair failed and we were unable to recover it. 00:26:55.691 [2024-07-15 18:12:49.136493] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.691 [2024-07-15 18:12:49.136510] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.136763] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.136779] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.136912] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.136928] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.137198] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.137239] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.137563] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.137595] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.137821] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.137853] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.138020] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.138051] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.138350] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.138367] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.138548] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.138563] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.138778] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.138809] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.139027] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.139058] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.139305] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.139322] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.139521] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.139537] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.139796] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.139826] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.140050] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.140082] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.140292] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.140325] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.140595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.140611] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.140808] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.140824] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.141082] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.141114] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.141323] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.141357] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.141600] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.141616] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.141904] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.141920] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.142183] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.142198] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.142482] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.142499] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.142651] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.142683] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.142831] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.142862] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.143090] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.143122] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.143325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.143341] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.143473] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.143489] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.143682] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.143714] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.143956] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.143987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.144164] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.144196] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.144520] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.144536] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.144714] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.144730] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.144926] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.692 [2024-07-15 18:12:49.144957] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.692 qpair failed and we were unable to recover it. 00:26:55.692 [2024-07-15 18:12:49.145247] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.145280] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.145457] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.145473] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.145657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.145674] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.145850] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.145881] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.146042] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.146074] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.146301] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.146334] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.146629] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.146645] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.146938] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.146970] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.147242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.147275] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.147491] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.147523] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.147809] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.147841] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.148009] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.148045] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.148350] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.148383] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.148601] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.148632] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.148848] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.148879] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.149087] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.149102] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.149376] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.149409] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.149635] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.149667] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.149949] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.149981] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.150294] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.150311] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.150547] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.150590] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.150802] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.150834] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.151135] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.151166] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.151444] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.151477] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.151730] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.151762] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.152075] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.152106] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.152387] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.152420] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.152665] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.152682] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.152930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.152962] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.153267] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.153299] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.153507] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.153523] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.153777] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.153793] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.153967] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.153982] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.154236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.154269] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.154594] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.154625] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.154923] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.154955] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.155262] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.155298] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.155585] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.155617] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.155800] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.155838] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.156049] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.156081] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.156268] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.156301] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.693 qpair failed and we were unable to recover it. 00:26:55.693 [2024-07-15 18:12:49.156509] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.693 [2024-07-15 18:12:49.156541] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.156747] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.156779] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.157027] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.157059] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.157200] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.157241] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.157541] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.157573] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.157740] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.157771] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.157995] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.158026] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.158302] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.158335] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.158657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.158689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.158921] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.158953] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.159300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.159332] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.159563] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.159580] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.159871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.159903] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.160182] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.160213] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.160471] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.160516] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.160779] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.160795] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.161013] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.161030] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.161315] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.161348] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.161577] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.161609] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.161910] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.161941] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.162173] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.162204] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.162442] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.162459] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.162657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.162689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.162990] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.163022] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.163318] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.163334] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.163533] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.163550] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.163763] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.163795] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.164080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.164111] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.164354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.164386] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.164621] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.164637] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.164908] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.694 [2024-07-15 18:12:49.164924] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.694 qpair failed and we were unable to recover it. 00:26:55.694 [2024-07-15 18:12:49.165183] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.165199] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.165420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.165436] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.165605] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.165636] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.165942] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.165974] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.166192] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.166208] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.166394] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.166411] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.166531] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.166546] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.166816] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.166891] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.167154] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.167189] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.167520] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.167554] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.167807] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.167840] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.168055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.168087] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.168323] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.168356] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.168570] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.168586] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.168862] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.168893] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.169117] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.169148] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.169376] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.169393] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.169639] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.169672] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.169950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.169982] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.170218] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.170261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.170499] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.170539] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.170764] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.170796] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.171036] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.171067] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.171307] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.171341] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.171640] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.171673] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.171910] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.171942] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.172248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.172281] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.172603] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.172635] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.172880] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.172912] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.173125] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.173158] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.173459] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.173492] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.173701] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.173734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.173969] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.174001] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.174311] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.174344] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.174581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.174613] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.174907] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.174940] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.175175] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.175207] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.175437] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.175478] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.175604] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.175621] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.175837] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.695 [2024-07-15 18:12:49.175854] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.695 qpair failed and we were unable to recover it. 00:26:55.695 [2024-07-15 18:12:49.176057] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.176089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.176379] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.176412] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.176623] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.176639] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.176755] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.176771] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.177038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.177055] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.177257] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.177274] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.177484] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.177500] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.177830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.177867] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.178125] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.178162] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.178407] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.178443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.178613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.178645] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.178893] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.178925] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.179205] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.179249] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.179552] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.179586] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.179833] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.179864] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.180121] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.180153] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.180336] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.180349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.180525] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.180556] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.180786] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.180818] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.181026] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.181056] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.181382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.181424] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.181668] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.181700] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.182035] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.182067] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.182210] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.182222] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.182474] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.182507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.182816] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.182849] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.183099] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.183131] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.183290] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.183324] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.183611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.183625] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.183746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.183759] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.184053] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.184085] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.184405] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.184418] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.184630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.184643] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.184833] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.184870] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.185122] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.185154] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.185349] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.185383] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.185660] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.185673] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.185798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.185811] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.185908] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.185919] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.186162] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.696 [2024-07-15 18:12:49.186175] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.696 qpair failed and we were unable to recover it. 00:26:55.696 [2024-07-15 18:12:49.186451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.697 [2024-07-15 18:12:49.186485] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.697 qpair failed and we were unable to recover it. 00:26:55.697 [2024-07-15 18:12:49.186794] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.697 [2024-07-15 18:12:49.186825] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.697 qpair failed and we were unable to recover it. 00:26:55.697 [2024-07-15 18:12:49.187145] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.697 [2024-07-15 18:12:49.187177] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.697 qpair failed and we were unable to recover it. 00:26:55.697 [2024-07-15 18:12:49.187504] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.697 [2024-07-15 18:12:49.187543] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.697 qpair failed and we were unable to recover it. 00:26:55.697 [2024-07-15 18:12:49.187844] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.697 [2024-07-15 18:12:49.187876] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.697 qpair failed and we were unable to recover it. 00:26:55.697 [2024-07-15 18:12:49.188174] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.697 [2024-07-15 18:12:49.188206] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.697 qpair failed and we were unable to recover it. 00:26:55.697 [2024-07-15 18:12:49.188506] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.697 [2024-07-15 18:12:49.188520] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.697 qpair failed and we were unable to recover it. 00:26:55.697 [2024-07-15 18:12:49.188740] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.697 [2024-07-15 18:12:49.188753] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.697 qpair failed and we were unable to recover it. 00:26:55.697 [2024-07-15 18:12:49.188943] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.697 [2024-07-15 18:12:49.188982] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.697 qpair failed and we were unable to recover it. 00:26:55.697 [2024-07-15 18:12:49.189149] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.189181] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.189491] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.189524] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.189755] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.189786] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.190094] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.190127] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.190424] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.190457] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.190651] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.190683] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.190902] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.190934] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.191179] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.191211] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.191478] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.191511] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.191737] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.191769] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.192005] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.192036] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.192267] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.192300] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.192619] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.192632] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.192810] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.192823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.193003] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.193015] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.193211] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.193261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.193496] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.193527] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.193814] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.193846] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.194165] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.194197] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.194442] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.194456] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.194670] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.194702] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.195001] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.195033] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.195281] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.195314] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.195480] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.195493] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.195686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.195718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.195959] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.195990] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.196277] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.196309] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.196594] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.196626] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.196949] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.196980] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.197287] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.197319] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.197519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.197551] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.197862] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.197893] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.198187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.198217] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.198512] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.198525] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.198721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.198734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.198914] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.198927] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.199204] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.199270] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.199382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.199396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.199591] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.199628] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.199932] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.199965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.698 qpair failed and we were unable to recover it. 00:26:55.698 [2024-07-15 18:12:49.200153] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.698 [2024-07-15 18:12:49.200185] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.200367] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.200400] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.200626] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.200657] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.200889] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.200922] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.201142] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.201174] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.201395] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.201408] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.201619] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.201651] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.201868] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.201901] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.202081] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.202112] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.202259] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.202272] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.202465] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.202479] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.202684] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.202715] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.202954] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.202986] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.203303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.203336] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.203688] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.203720] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.204008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.204040] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.204352] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.204366] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.204564] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.204578] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.204849] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.204863] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.205078] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.205091] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.205200] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.205211] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.205442] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.205475] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.205715] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.205747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.205893] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.205926] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.206244] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.206277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.206519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.206551] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.206843] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.206856] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.207067] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.207081] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.207329] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.207363] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.207602] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.207634] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.207922] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.207953] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.208173] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.208204] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.208439] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.208471] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.208746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.208780] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.209004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.209036] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.209248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.209282] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.209561] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.209574] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.209821] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.209853] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.210136] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.210174] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.210484] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.210518] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.210780] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.699 [2024-07-15 18:12:49.210811] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.699 qpair failed and we were unable to recover it. 00:26:55.699 [2024-07-15 18:12:49.211182] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.211213] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.211451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.211464] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.211672] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.211705] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.211989] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.212021] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.212354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.212387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.212628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.212660] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.212908] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.212940] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.213258] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.213291] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.213508] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.213522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.213625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.213666] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.213981] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.214013] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.214165] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.214197] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.214480] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.214516] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.214800] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.214832] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.215051] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.215083] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.215387] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.215401] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.215589] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.215602] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.215796] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.215828] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.216042] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.216074] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.216304] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.216338] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.216652] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.216684] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.216990] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.217028] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.217290] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.217323] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.217555] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.217587] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.217936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.217969] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.218252] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.218284] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.218619] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.218652] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.218868] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.218900] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.219065] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.219097] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.219432] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.219465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.219644] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.219676] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.219957] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.219990] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.220216] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.220261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.220509] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.220549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.220736] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.220749] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.220936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.220960] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.221143] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.221175] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.221432] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.221472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.221641] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.221654] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.221849] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.221882] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.700 [2024-07-15 18:12:49.222021] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.700 [2024-07-15 18:12:49.222053] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.700 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.222366] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.222400] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.222729] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.222762] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.222951] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.222983] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.223303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.223317] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.223424] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.223448] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.223642] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.223655] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.223839] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.223871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.224095] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.224126] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.224475] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.224508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.224723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.224755] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.224981] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.225013] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.225297] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.225329] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.225555] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.225587] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.225874] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.225908] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.226140] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.226172] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.226487] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.226501] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.226681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.226713] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.227018] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.227050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.227391] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.227404] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.227648] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.227662] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.227929] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.227942] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.228071] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.228084] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.228356] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.228390] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.228682] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.228714] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.228953] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.228985] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.229213] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.229257] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.229471] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.229484] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.229681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.229713] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.229930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.229962] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.230186] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.230218] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.230414] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.230446] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.230730] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.230761] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.230990] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.231022] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.231384] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.231417] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.231706] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.231738] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.231975] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.232007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.232238] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.232277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.232499] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.232531] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.232743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.232775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.232987] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.233019] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.701 [2024-07-15 18:12:49.233245] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.701 [2024-07-15 18:12:49.233278] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.701 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.233504] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.233536] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.233743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.233775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.233996] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.234028] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.234283] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.234317] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.234537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.234550] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.234741] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.234754] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.234954] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.234967] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.235193] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.235207] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.235492] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.235506] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.235775] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.235807] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.236028] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.236060] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.236371] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.236412] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.236586] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.236599] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.236837] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.236850] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.236993] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.237027] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.237261] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.237294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.237516] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.237562] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.237855] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.237887] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.238048] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.238080] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.238391] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.238404] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.238596] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.238629] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.238964] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.238996] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.239262] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.239295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.239506] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.239519] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.239768] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.239800] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.240129] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.240162] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.240311] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.240360] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.240574] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.240606] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.240933] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.240965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.241278] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.241311] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.241592] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.241624] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.241936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.241969] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.242296] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.242329] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.242509] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.242541] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.242758] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.702 [2024-07-15 18:12:49.242789] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.702 qpair failed and we were unable to recover it. 00:26:55.702 [2024-07-15 18:12:49.243006] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.243044] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.243262] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.243296] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.243616] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.243648] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.243933] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.243965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.244266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.244300] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.244529] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.244562] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.244726] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.244759] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.245048] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.245080] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.245370] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.245403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.245718] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.245750] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.246059] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.246091] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.246328] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.246361] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.246612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.246644] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.246861] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.246893] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.247149] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.247182] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.247468] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.247502] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.247817] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.247849] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.248164] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.248197] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.248455] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.248468] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.248682] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.248715] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.248952] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.248984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.249294] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.249327] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.249566] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.249598] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.249833] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.249864] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.250077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.250109] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.250340] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.250373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.250600] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.250632] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.250837] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.250850] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.251126] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.251157] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.251407] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.251440] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.251675] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.251707] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.251862] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.251895] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.252206] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.252249] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.252533] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.252578] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.252826] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.252858] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.253096] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.253129] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.253438] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.253475] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.253589] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.253603] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.253746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.253779] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.254084] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.254117] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.703 [2024-07-15 18:12:49.254418] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.703 [2024-07-15 18:12:49.254456] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.703 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.254757] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.254789] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.255023] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.255055] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.255364] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.255396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.255588] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.255601] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.255778] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.255810] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.256068] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.256101] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.256394] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.256426] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.256720] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.256752] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.257085] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.257117] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.257429] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.257461] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.257756] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.257789] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.258084] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.258097] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.258431] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.258463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.258704] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.258737] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.258977] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.259008] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.259177] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.259210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.259451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.259484] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.259647] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.259675] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.259980] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.260012] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.260256] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.260289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.260594] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.260607] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.260827] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.260859] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.261171] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.261203] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.261501] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.261534] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.261842] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.261874] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.262169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.262201] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.262436] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.262469] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.262650] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.262682] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.262991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.263004] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.263195] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.263208] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.263409] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.263422] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.263689] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.263701] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.263898] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.263912] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.264175] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.264188] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.264408] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.264422] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.264719] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.264760] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.264991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.265023] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.265332] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.265366] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.265630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.265665] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.265991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.266028] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.704 [2024-07-15 18:12:49.266294] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.704 [2024-07-15 18:12:49.266328] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.704 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.266577] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.266590] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.266838] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.266852] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.267055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.267087] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.267403] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.267437] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.267746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.267778] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.268009] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.268041] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.268348] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.268389] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.268637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.268650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.268851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.268864] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.269104] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.269117] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.269381] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.269395] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.269653] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.269686] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.269946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.269979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.270307] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.270340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.270554] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.270568] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.270819] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.270851] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.271167] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.271199] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.271532] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.271566] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.271781] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.271814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.272105] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.272119] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.272415] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.272449] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.272629] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.272642] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.272890] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.272922] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.273160] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.273192] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.273550] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.273583] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.273837] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.273870] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.274155] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.274187] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.274442] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.274456] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.274725] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.274757] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.274974] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.275006] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.275244] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.275277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.275509] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.275540] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.275845] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.275878] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.276192] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.276235] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.276561] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.276593] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.276911] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.276942] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.277246] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.277278] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.277576] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.277608] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.277910] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.277948] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.278182] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.278214] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.705 [2024-07-15 18:12:49.278569] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.705 [2024-07-15 18:12:49.278601] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.705 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.278901] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.278914] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.279184] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.279198] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.279373] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.279386] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.279597] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.279628] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.279939] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.279971] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.280187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.280219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.280484] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.280517] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.280686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.280718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.280913] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.280927] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.281192] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.281206] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.281459] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.281492] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.281755] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.281788] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.282086] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.282111] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.282425] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.282458] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.282619] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.282651] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.282898] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.282929] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.283259] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.283292] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.283604] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.283637] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.283921] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.283953] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.284247] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.284279] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.284494] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.284525] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.284864] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.284895] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.285147] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.285180] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.285524] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.285558] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.285776] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.285809] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.286067] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.286099] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.286390] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.286423] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.286657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.286670] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.286915] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.286949] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.287266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.287299] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.287551] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.287565] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.287742] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.287773] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.288023] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.288055] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.288292] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.288325] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.288675] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.288706] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.706 qpair failed and we were unable to recover it. 00:26:55.706 [2024-07-15 18:12:49.288916] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.706 [2024-07-15 18:12:49.288948] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.289259] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.289292] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.289531] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.289568] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.289917] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.289949] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.290256] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.290289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.290608] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.290640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.290927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.290959] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.291264] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.291298] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.291601] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.291633] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.291939] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.291971] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.292203] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.292243] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.292514] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.292527] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.292774] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.292806] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.293055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.293088] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.293399] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.293432] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.293659] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.293671] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.293862] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.293876] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.294072] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.294103] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.294317] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.294350] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.294663] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.294696] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.294927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.294941] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.295189] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.295221] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.295541] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.295573] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.295786] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.295830] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.296022] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.296035] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.296318] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.296332] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.296612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.296643] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.296978] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.297009] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.297320] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.297353] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.297647] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.297680] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.297989] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.298021] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.298243] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.298277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.298585] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.298617] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.298923] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.298957] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.299255] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.299289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.299503] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.299535] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.299842] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.299855] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.300110] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.300123] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.300315] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.300329] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.300624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.300656] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.707 [2024-07-15 18:12:49.300970] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.707 [2024-07-15 18:12:49.301002] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.707 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.301221] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.301263] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.301496] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.301534] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.301835] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.301868] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.302027] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.302060] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.302341] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.302374] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.302686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.302718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.302997] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.303030] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.303339] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.303373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.303671] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.303703] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.304006] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.304019] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.304204] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.304217] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.304414] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.304447] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.304758] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.304791] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.305011] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.305043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.305339] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.305373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.305666] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.305680] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.305999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.306031] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.306177] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.306209] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.306432] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.306465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.306751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.306792] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.307101] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.307133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.307444] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.307478] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.307788] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.307821] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.308126] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.308159] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.308397] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.308431] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.308596] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.308628] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.308928] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.308961] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.309262] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.309296] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.309543] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.309576] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.309927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.309959] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.310178] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.310210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.310486] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.310519] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.310734] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.310766] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.310984] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.311016] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.311331] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.311364] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.311549] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.311582] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.311867] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.311900] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.312140] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.312172] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.312498] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.708 [2024-07-15 18:12:49.312532] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.708 qpair failed and we were unable to recover it. 00:26:55.708 [2024-07-15 18:12:49.312819] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.312851] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.313169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.313183] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.313486] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.313526] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.313755] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.313787] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.313950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.313982] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.314267] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.314300] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.314583] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.314617] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.314905] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.314938] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.315107] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.315140] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.315368] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.315401] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.315733] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.315766] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.315984] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.316016] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.316332] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.316365] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.316651] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.316685] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.316999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.317011] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.317305] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.317319] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.317450] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.317464] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.317732] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.317764] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.318101] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.318133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.318442] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.318475] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.318700] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.318713] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.318907] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.318920] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.319129] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.319142] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.319327] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.319342] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.319517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.319550] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.319833] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.319866] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.320120] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.320153] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.320372] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.320405] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.320697] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.320729] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.321031] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.321044] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.321324] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.321358] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.321701] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.321734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.322013] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.322038] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.322170] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.322202] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.322464] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.322496] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.322830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.322863] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.323116] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.323130] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.323377] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.323411] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.323629] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.323661] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.323991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.324023] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.324275] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.709 [2024-07-15 18:12:49.324308] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.709 qpair failed and we were unable to recover it. 00:26:55.709 [2024-07-15 18:12:49.324591] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.324622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.324928] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.324966] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.325261] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.325295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.325617] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.325649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.325978] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.326009] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.326245] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.326278] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.326630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.326662] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.326920] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.326953] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.327249] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.327282] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.327588] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.327620] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.327848] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.327880] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.328188] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.328220] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.328509] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.328552] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.328846] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.328878] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.329161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.329193] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.329425] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.329460] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.329724] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.329737] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.330016] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.330048] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.330358] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.330392] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.330635] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.330649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.330833] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.330846] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.331045] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.331077] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.331336] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.331368] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.331660] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.331693] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.332001] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.332033] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.332250] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.332284] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.332515] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.332547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.332878] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.332903] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.333214] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.333255] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.333551] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.333582] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.333796] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.333828] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.333974] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.334006] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.334184] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.334216] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.334536] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.334569] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.334795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.334827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.335130] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.710 [2024-07-15 18:12:49.335162] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.710 qpair failed and we were unable to recover it. 00:26:55.710 [2024-07-15 18:12:49.335394] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.335428] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.335643] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.335676] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.335961] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.335993] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.336257] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.336290] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.336470] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.336503] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.336738] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.336771] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.337121] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.337154] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.337379] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.337412] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.337630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.337661] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.337980] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.337993] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.338181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.338195] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.338392] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.338424] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.338664] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.338696] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.338988] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.339020] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.339333] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.339367] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.339657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.339689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.340002] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.340034] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.340262] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.340295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.340584] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.340616] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.340894] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.340924] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.341255] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.341287] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.341596] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.341630] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.341918] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.341949] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.342264] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.342297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.342607] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.342639] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.342853] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.342885] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.343166] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.343198] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.343521] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.343553] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.343829] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.343864] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.344100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.344133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.344466] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.344500] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.344783] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.344814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.345142] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.345179] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.345499] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.345533] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.345827] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.345840] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.346152] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.346184] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.346503] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.346536] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.346771] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.346784] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.346961] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.346974] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.347251] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.711 [2024-07-15 18:12:49.347284] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.711 qpair failed and we were unable to recover it. 00:26:55.711 [2024-07-15 18:12:49.347518] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.347550] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.347863] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.347896] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.348204] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.348269] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.348496] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.348529] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.348880] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.348911] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.349145] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.349178] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.349541] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.349575] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.349879] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.349911] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.350213] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.350257] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.350482] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.350514] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.350726] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.350739] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.350941] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.350974] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.351256] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.351289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.351597] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.351629] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.351868] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.351901] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.352213] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.352269] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.352420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.352452] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.352685] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.352717] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.353028] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.353060] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.353300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.353333] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.353682] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.353714] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.353967] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.353999] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.354238] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.354271] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.354563] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.354595] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.354774] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.354806] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.354977] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.354990] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.355183] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.355215] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.355541] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.355573] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.355789] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.355821] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.356113] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.356126] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.356312] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.356326] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.356541] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.356573] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.356789] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.356827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.357055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.357068] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.357265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.357299] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.357513] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.357545] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.357768] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.357781] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.358026] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.358057] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.358202] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.358245] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.358532] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.358565] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.712 qpair failed and we were unable to recover it. 00:26:55.712 [2024-07-15 18:12:49.358809] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.712 [2024-07-15 18:12:49.358822] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.359054] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.359068] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.359367] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.359401] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.359654] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.359686] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.360020] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.360052] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.360384] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.360417] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.360710] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.360742] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.361076] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.361108] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.361425] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.361458] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.361763] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.361795] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.362083] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.362120] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.362351] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.362386] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/host/target_disconnect.sh: line 36: 745985 Killed "${NVMF_APP[@]}" "$@" 00:26:55.713 [2024-07-15 18:12:49.362624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.362657] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.362908] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.362922] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 18:12:49 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@48 -- # disconnect_init 10.0.0.2 00:26:55.713 [2024-07-15 18:12:49.363214] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.363244] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.363378] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.363393] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 18:12:49 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@17 -- # nvmfappstart -m 0xF0 00:26:55.713 [2024-07-15 18:12:49.363637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.363650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 18:12:49 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:26:55.713 [2024-07-15 18:12:49.363825] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.363839] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 18:12:49 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@722 -- # xtrace_disable 00:26:55.713 [2024-07-15 18:12:49.364112] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.364127] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.364308] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.364322] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 18:12:49 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:55.713 [2024-07-15 18:12:49.364534] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.364548] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.364838] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.364852] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.365053] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.365067] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.365288] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.365302] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.365498] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.365511] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.365618] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.365629] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.365734] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.365745] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.365874] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.365887] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.366178] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.366191] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.366432] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.366445] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.366585] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.366599] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.366770] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.366784] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.367028] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.367043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.367310] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.367325] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.367508] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.367522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.367644] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.367655] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.367855] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.367867] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.368116] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.368126] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.713 [2024-07-15 18:12:49.368327] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.713 [2024-07-15 18:12:49.368340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.713 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.368553] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.368567] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.368716] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.368730] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.368908] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.368921] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.369167] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.369180] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.369411] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.369424] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.369610] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.369625] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.369871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.369884] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.370151] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.370164] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.370419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.370433] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.370576] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.370589] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.370762] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.370776] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.371020] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.371033] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 18:12:49 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@481 -- # nvmfpid=746783 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.371247] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.371261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 18:12:49 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@482 -- # waitforlisten 746783 00:26:55.714 18:12:49 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF0 00:26:55.714 [2024-07-15 18:12:49.371551] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.371566] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 18:12:49 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@829 -- # '[' -z 746783 ']' 00:26:55.714 [2024-07-15 18:12:49.371807] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.371822] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.372016] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 18:12:49 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:26:55.714 [2024-07-15 18:12:49.372030] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.372239] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.372252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 18:12:49 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@834 -- # local max_retries=100 00:26:55.714 [2024-07-15 18:12:49.372472] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.372486] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 18:12:49 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:26:55.714 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:26:55.714 [2024-07-15 18:12:49.372681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.372696] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 18:12:49 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@838 -- # xtrace_disable 00:26:55.714 [2024-07-15 18:12:49.372960] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.372975] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 18:12:49 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:55.714 [2024-07-15 18:12:49.373114] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.373129] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.373303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.373316] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.373495] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.373508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.373751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.373766] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.373882] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.373895] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.374113] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.374127] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.374405] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.374421] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.374689] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.374704] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.374977] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.374990] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.375169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.375183] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.375385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.375401] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.375669] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.375681] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.375862] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.714 [2024-07-15 18:12:49.375875] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.714 qpair failed and we were unable to recover it. 00:26:55.714 [2024-07-15 18:12:49.376139] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.376151] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.376408] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.376422] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.376649] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.376663] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.376785] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.376799] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.376998] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.377010] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.377205] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.377217] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.377467] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.377480] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.377600] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.377614] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.377810] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.377823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.378017] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.378031] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.378263] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.378277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.378508] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.378521] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.378766] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.378779] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.378892] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.378904] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.379097] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.379111] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.379381] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.379394] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.379588] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.379601] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.379800] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.379814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.380022] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.380035] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.380282] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.380296] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.380488] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.380502] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.380751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.380765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.380950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.380963] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.381107] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.381120] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.381321] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.381335] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.381608] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.381621] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.381737] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.381748] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.381932] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.381945] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.382153] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.382166] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.382407] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.382420] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.382615] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.382628] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.382921] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.382934] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.383212] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.383231] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.383404] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.383419] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.383698] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.383712] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.383829] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.383842] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.715 [2024-07-15 18:12:49.384033] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.715 [2024-07-15 18:12:49.384046] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.715 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.384220] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.384242] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.384485] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.384498] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.384670] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.384684] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.384850] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.384864] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.385059] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.385072] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.385250] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.385264] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.385459] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.385472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.385722] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.385736] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.386026] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.386040] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.386210] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.386230] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.386420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.386433] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.386695] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.386709] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.386930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.386944] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.387120] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.387133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.387250] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.387262] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.387455] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.387468] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.387661] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.387674] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.387924] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.387937] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.388143] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.388156] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.388460] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.388473] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.388667] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.388680] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.388922] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.388935] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.389071] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.389085] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.389269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.389282] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.389460] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.389473] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.389662] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.389675] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.389935] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.389947] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.390148] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.390161] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.390365] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.390378] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.390498] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.390511] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.390669] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.390682] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.390922] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.390935] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.391112] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.391125] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.391394] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.391408] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.391711] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.391724] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.392010] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.392023] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.392135] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.392148] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.392380] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.392393] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.392646] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.716 [2024-07-15 18:12:49.392658] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.716 qpair failed and we were unable to recover it. 00:26:55.716 [2024-07-15 18:12:49.392935] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.717 [2024-07-15 18:12:49.392948] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.717 qpair failed and we were unable to recover it. 00:26:55.717 [2024-07-15 18:12:49.393119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.717 [2024-07-15 18:12:49.393132] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.717 qpair failed and we were unable to recover it. 00:26:55.717 [2024-07-15 18:12:49.393376] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.717 [2024-07-15 18:12:49.393389] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.717 qpair failed and we were unable to recover it. 00:26:55.717 [2024-07-15 18:12:49.393635] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.717 [2024-07-15 18:12:49.393647] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.717 qpair failed and we were unable to recover it. 00:26:55.717 [2024-07-15 18:12:49.393823] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.717 [2024-07-15 18:12:49.393835] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.717 qpair failed and we were unable to recover it. 00:26:55.717 [2024-07-15 18:12:49.394027] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.717 [2024-07-15 18:12:49.394040] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.717 qpair failed and we were unable to recover it. 00:26:55.717 [2024-07-15 18:12:49.394237] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.717 [2024-07-15 18:12:49.394251] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.717 qpair failed and we were unable to recover it. 00:26:55.717 [2024-07-15 18:12:49.394418] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.717 [2024-07-15 18:12:49.394431] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.717 qpair failed and we were unable to recover it. 00:26:55.717 [2024-07-15 18:12:49.394708] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.717 [2024-07-15 18:12:49.394721] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.717 qpair failed and we were unable to recover it. 00:26:55.717 [2024-07-15 18:12:49.394910] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.717 [2024-07-15 18:12:49.394924] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.717 qpair failed and we were unable to recover it. 00:26:55.717 [2024-07-15 18:12:49.395240] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.717 [2024-07-15 18:12:49.395253] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.717 qpair failed and we were unable to recover it. 00:26:55.717 [2024-07-15 18:12:49.395506] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.717 [2024-07-15 18:12:49.395519] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.717 qpair failed and we were unable to recover it. 00:26:55.717 [2024-07-15 18:12:49.395754] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.717 [2024-07-15 18:12:49.395767] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.717 qpair failed and we were unable to recover it. 00:26:55.717 [2024-07-15 18:12:49.395985] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.717 [2024-07-15 18:12:49.395998] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.717 qpair failed and we were unable to recover it. 00:26:55.717 [2024-07-15 18:12:49.396194] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.717 [2024-07-15 18:12:49.396208] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.717 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.396403] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.396417] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.396542] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.396557] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.396803] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.396818] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.397006] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.397020] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.397148] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.397160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.397336] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.397350] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.397457] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.397468] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.397699] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.397712] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.397887] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.397900] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.398087] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.398100] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.398258] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.398273] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.398545] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.398558] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.398752] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.398765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.398941] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.398954] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.399197] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.399210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.399424] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.399438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.399626] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.399639] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.399826] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.399839] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.399978] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.399991] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.400131] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.400143] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.400333] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.400346] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.400641] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.400654] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.400934] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.400950] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.401148] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.401160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.401422] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.401435] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.401743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.401756] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.402001] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.402014] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.402257] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.402270] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.402456] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.402470] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.402672] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.402684] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.402879] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.402892] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.403030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.403043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.403282] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.403294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.997 [2024-07-15 18:12:49.403495] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.997 [2024-07-15 18:12:49.403508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.997 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.403679] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.403692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.403944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.403957] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.404060] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.404072] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.404269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.404282] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.404541] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.404554] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.404845] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.404858] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.405029] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.405042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.405245] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.405257] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.405455] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.405468] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.405729] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.405741] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.405931] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.405944] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.406065] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.406078] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.406265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.406278] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.406539] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.406552] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.406742] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.406754] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.407045] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.407057] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.407251] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.407264] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.407440] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.407452] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.407712] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.407725] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.407942] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.407954] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.408142] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.408155] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.408260] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.408272] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.408399] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.408412] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.408540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.408554] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.408664] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.408678] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.408939] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.408952] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.409209] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.409221] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.409400] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.409413] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.409555] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.409571] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.409808] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.409821] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.410008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.410021] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.410181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.410193] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.410434] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.410447] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.410709] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.410722] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.410907] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.410920] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.411156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.411168] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.411427] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.411439] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.411554] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.411566] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.411748] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.998 [2024-07-15 18:12:49.411761] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.998 qpair failed and we were unable to recover it. 00:26:55.998 [2024-07-15 18:12:49.411879] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.411891] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.412074] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.412087] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.412289] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.412302] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.412486] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.412498] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.412757] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.412769] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.413002] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.413015] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.413275] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.413288] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.413432] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.413443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.413580] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.413593] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.413804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.413817] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.413948] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.413960] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.414163] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.414176] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.414345] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.414357] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.414552] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.414565] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.414798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.414811] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.415008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.415020] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.415215] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.415234] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.415471] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.415483] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.415743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.415756] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.415924] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.415936] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.416136] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.416148] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.416287] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.416299] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.416480] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.416493] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.416743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.416756] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.416993] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.417006] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.417166] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:26:55.999 [2024-07-15 18:12:49.417201] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.417212] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 [2024-07-15 18:12:49.417212] [ DPDK EAL parameters: nvmf -c 0xF0 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.417404] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.417416] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.417671] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.417681] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.417865] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.417877] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.417999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.418009] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.418188] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.418199] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.418451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.418462] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.418631] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.418644] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.418768] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.418779] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.419015] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.419027] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.419211] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.419229] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.419351] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.419363] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.419534] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:55.999 [2024-07-15 18:12:49.419547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:55.999 qpair failed and we were unable to recover it. 00:26:55.999 [2024-07-15 18:12:49.419753] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.419766] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.420013] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.420026] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.420216] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.420240] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.420422] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.420435] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.420677] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.420690] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.420892] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.420905] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.421167] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.421180] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.421360] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.421373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.421633] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.421646] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.421909] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.421922] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.422100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.422112] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.422371] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.422384] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.422508] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.422521] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.422801] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.422813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.423047] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.423059] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.423234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.423247] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.423432] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.423444] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.423671] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.423684] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.423940] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.423952] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.424135] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.424148] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.424352] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.424365] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.424535] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.424547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.424801] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.424813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.425066] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.425078] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.425177] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.425189] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.425355] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.425367] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.425631] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.425643] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.425878] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.425890] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.426142] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.426155] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.426343] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.426357] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.426533] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.426548] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.426667] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.426679] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.426936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.426948] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.427137] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.427149] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.427408] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.427421] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.427604] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.427617] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.427830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.427842] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.428124] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.428136] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.428369] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.428382] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.428620] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.428632] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.000 [2024-07-15 18:12:49.428909] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.000 [2024-07-15 18:12:49.428922] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.000 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.429111] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.429122] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.429313] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.429326] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.429511] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.429524] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.429692] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.429704] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.429870] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.429883] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.430076] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.430089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.430325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.430338] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.430572] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.430585] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.430852] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.430865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.431030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.431042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.431236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.431249] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.431452] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.431465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.431644] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.431656] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.431907] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.431920] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.432174] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.432187] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.432438] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.432451] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.432697] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.432710] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.432923] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.432936] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.433118] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.433131] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.433409] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.433422] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.433620] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.001 [2024-07-15 18:12:49.433632] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.001 qpair failed and we were unable to recover it. 00:26:56.001 [2024-07-15 18:12:49.433757] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.433769] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.433933] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.433945] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.434167] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.434179] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.434422] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.434435] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.434532] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.434545] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.434802] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.434814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.435078] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.435091] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.435272] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.435285] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.435469] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.435484] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.435599] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.435611] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.435840] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.435852] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.436027] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.436040] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.436161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.436173] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.436430] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.436443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.436609] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.436622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.436878] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.436891] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.437125] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.437138] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.437371] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.437384] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.437569] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.437582] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.437838] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.437850] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.438040] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.438052] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.438297] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.438309] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.438571] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.438584] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.438827] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.438839] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.438951] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.438964] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.439230] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.439244] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.439418] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.439431] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.002 [2024-07-15 18:12:49.439616] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.002 [2024-07-15 18:12:49.439628] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.002 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.439909] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.439922] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.440197] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.440210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.440407] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.440420] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.440610] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.440622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.440866] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.440879] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.441108] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.441121] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.441382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.441395] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.441563] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.441576] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.441832] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.441845] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.442105] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.442117] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.442302] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.442316] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.442519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.442532] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.442698] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.442710] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.442876] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.442889] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.443071] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.443083] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.443248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.443261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.443437] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.443450] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.443566] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.443578] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.443692] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.443704] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.443946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.443959] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.444151] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.444165] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.444396] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.444409] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.444525] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.444537] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.444796] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.444808] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 EAL: No free 2048 kB hugepages reported on node 1 00:26:56.003 [2024-07-15 18:12:49.445043] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.445056] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.445313] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.445326] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.445454] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.003 [2024-07-15 18:12:49.445467] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.003 qpair failed and we were unable to recover it. 00:26:56.003 [2024-07-15 18:12:49.445641] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.445654] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.445911] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.445924] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.446161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.446173] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.446359] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.446372] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.446628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.446640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.446894] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.446906] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.447140] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.447152] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.447318] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.447330] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.447549] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.447562] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.447805] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.447818] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.447927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.447939] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.448169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.448181] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.448397] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.448410] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.448666] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.448679] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.448862] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.448874] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.449052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.449064] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.449231] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.449244] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.449444] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.449456] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.449630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.449642] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.449901] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.449913] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.450117] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.450128] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.450264] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.450277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.450528] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.450540] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.450745] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.450757] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.450952] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.450964] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.451127] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.451140] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.451324] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.451336] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.004 qpair failed and we were unable to recover it. 00:26:56.004 [2024-07-15 18:12:49.451521] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.004 [2024-07-15 18:12:49.451534] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.451635] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.451647] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.451829] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.451841] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.452036] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.452048] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.452222] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.452249] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.452493] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.452505] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.452691] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.452705] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.452944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.452956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.453189] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.453202] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.453479] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.453492] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.453667] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.453679] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.453933] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.453946] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.454129] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.454141] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.454323] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.454335] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.454616] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.454629] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.454820] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.454833] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.454944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.454956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.455052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.455064] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.455303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.455315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.455483] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.455495] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.455730] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.455742] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.455972] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.455984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.456164] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.456176] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.456426] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.456438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.456692] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.456704] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.456968] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.456981] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.457265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.457277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.457533] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.005 [2024-07-15 18:12:49.457546] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.005 qpair failed and we were unable to recover it. 00:26:56.005 [2024-07-15 18:12:49.457804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.457816] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.458050] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.458062] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.458242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.458255] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.458513] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.458526] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.458710] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.458722] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.458891] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.458904] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.459136] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.459148] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.459398] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.459411] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.459532] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.459544] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.459798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.459810] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.459944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.459956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.460088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.460100] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.460197] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.460209] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.460478] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.460491] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.460754] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.460766] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.460952] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.460964] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.461220] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.461235] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.461403] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.461416] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.461615] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.461629] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.461834] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.461847] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.462119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.462131] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.462407] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.462420] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.462630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.462643] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.462827] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.462838] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.462957] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.462969] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.463223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.463241] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.463495] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.463507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.006 [2024-07-15 18:12:49.463703] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.006 [2024-07-15 18:12:49.463716] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.006 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.463997] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.464009] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.464263] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.464276] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.464531] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.464543] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.464722] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.464734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.464966] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.464978] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.465164] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.465176] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.465414] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.465426] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.465682] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.465695] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.465952] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.465965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.466129] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.466140] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.466322] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.466335] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.466597] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.466610] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.466774] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.466786] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.467013] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.467025] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.467186] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.467199] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.467439] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.467452] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.467555] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.467566] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.467739] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.467751] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.468030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.468041] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.468290] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.468302] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.468466] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.468479] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.468725] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.468737] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.468991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.469003] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.469176] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.469188] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.469348] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.469362] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.007 qpair failed and we were unable to recover it. 00:26:56.007 [2024-07-15 18:12:49.469540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.007 [2024-07-15 18:12:49.469553] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.469809] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.469820] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.469998] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.470011] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.470262] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.470275] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.470527] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.470540] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.470769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.470783] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.470974] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.470986] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.471165] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.471177] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.471430] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.471443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.471633] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.471645] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.471855] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.471868] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.471977] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.471989] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.472179] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.472191] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.472420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.472433] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.472600] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.472612] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.472865] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.472878] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.472991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.473003] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.473252] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.473264] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.473430] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.473442] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.473687] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.473699] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.473879] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.473890] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.474117] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.474129] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.474342] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.474355] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.474676] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.008 [2024-07-15 18:12:49.474688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.008 qpair failed and we were unable to recover it. 00:26:56.008 [2024-07-15 18:12:49.474860] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.474872] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.474965] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.474977] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.475142] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.475154] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.475334] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.475346] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.475571] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.475584] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.475765] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.475776] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.475948] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.475960] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.476220] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.476237] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.476490] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.476502] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.476681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.476692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.476870] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.476881] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.476979] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.476991] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.477169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.477182] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.477369] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.477381] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.477485] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.477497] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.477703] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.477714] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.477961] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.477973] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.478229] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.478241] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.478481] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.478493] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.478732] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.478744] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.478922] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.478933] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.479186] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.479200] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.479430] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.479442] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.479617] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.479628] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.479878] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.479889] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.480145] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.480157] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.480326] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.480339] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.480449] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.009 [2024-07-15 18:12:49.480461] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.009 qpair failed and we were unable to recover it. 00:26:56.009 [2024-07-15 18:12:49.480709] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.480721] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.480984] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.480997] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.481121] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.481133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.481366] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.481378] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.481606] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.481619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.481734] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.481746] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.481983] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.481995] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.482107] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.482119] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.482241] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.482252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.482428] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.482440] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.482639] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.482652] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.482827] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.482839] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.482950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.482961] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.483052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.483063] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.483250] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.483262] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.483511] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.483523] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.483731] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.483743] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.483936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.483948] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.484190] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.484201] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.484453] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.484465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.484661] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.484673] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.484855] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.484867] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.485043] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.485055] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.485282] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.485294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.485521] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.485533] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.485770] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.485782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.485955] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.485967] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.486078] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.010 [2024-07-15 18:12:49.486090] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.010 qpair failed and we were unable to recover it. 00:26:56.010 [2024-07-15 18:12:49.486288] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.486300] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.486480] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.486492] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.486672] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.486684] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.486885] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.486897] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.487059] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.487071] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.487299] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.487313] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.487544] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.487557] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.487831] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.487843] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.488080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.488092] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.488371] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.488383] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.488491] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.488503] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.488741] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.488754] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.489003] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.489016] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.489266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.489278] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.489476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.489488] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.489634] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:26:56.011 [2024-07-15 18:12:49.489686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.489698] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.489820] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.489832] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.490088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.490101] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.490346] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.490360] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.490610] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.490622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.490901] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.490914] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.491077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.491090] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.491339] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.491352] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.491524] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.491537] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.491837] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.491850] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.492060] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.492073] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.492346] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.011 [2024-07-15 18:12:49.492359] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.011 qpair failed and we were unable to recover it. 00:26:56.011 [2024-07-15 18:12:49.492544] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.492556] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.492668] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.492680] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.492786] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.492798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.492978] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.492990] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.493101] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.493113] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.493341] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.493354] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.493541] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.493553] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.493804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.493816] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.493984] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.493996] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.494246] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.494258] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.494422] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.494435] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.494699] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.494713] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.494908] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.494920] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.495210] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.495222] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.495477] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.495490] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.495652] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.495666] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.495863] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.495875] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.496134] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.496146] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.496332] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.496345] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.496534] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.496547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.496731] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.496743] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.496992] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.497004] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.497169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.497182] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.497385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.497398] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.497627] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.497640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.497893] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.497906] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.498088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.498101] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.498355] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.012 [2024-07-15 18:12:49.498368] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.012 qpair failed and we were unable to recover it. 00:26:56.012 [2024-07-15 18:12:49.498621] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.498634] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.498839] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.498851] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.499045] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.499058] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.499239] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.499256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.499499] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.499512] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.499747] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.499760] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.499939] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.499952] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.500203] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.500216] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.500473] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.500485] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.500723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.500734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.500912] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.500924] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.501171] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.501183] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.501410] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.501423] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.501606] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.501618] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.501713] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.501724] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.501975] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.501987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.502195] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.502206] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.502454] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.502467] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.502649] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.502660] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.502908] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.502919] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.503144] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.503157] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.503405] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.503416] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.503652] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.503665] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.503840] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.503852] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.504078] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.504089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.504304] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.504316] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.504542] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.504554] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.013 [2024-07-15 18:12:49.504675] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.013 [2024-07-15 18:12:49.504686] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.013 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.504863] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.504875] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.014 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.505109] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.505121] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.014 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.505322] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.505335] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.014 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.505587] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.505599] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.014 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.505691] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.505703] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.014 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.505894] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.505906] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.014 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.506098] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.506110] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.014 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.506359] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.506371] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.014 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.506534] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.506545] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.014 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.506725] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.506738] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.014 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.507016] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.507027] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.014 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.507187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.507199] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.014 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.507360] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.507372] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.014 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.507622] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.507634] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.014 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.507810] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.507821] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.014 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.508050] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.508064] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.014 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.508311] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.508322] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.014 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.508483] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.508495] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.014 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.508688] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.508700] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.014 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.508801] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.508812] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.014 qpair failed and we were unable to recover it. 00:26:56.014 [2024-07-15 18:12:49.509015] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.014 [2024-07-15 18:12:49.509027] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.509214] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.509230] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.509349] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.509360] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.509634] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.509646] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.509894] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.509906] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.510083] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.510094] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.510265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.510277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.510438] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.510450] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.510701] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.510713] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.510890] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.510902] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.511155] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.511166] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.511342] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.511354] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.511542] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.511554] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.511780] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.511792] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.511956] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.511968] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.512142] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.512154] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.512382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.512394] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.512647] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.512658] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.512855] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.512867] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.512990] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.513002] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.513244] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.513256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.513364] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.513376] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.513634] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.513646] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.513843] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.513854] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.514048] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.514059] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.514234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.514246] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.514427] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.514439] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.514615] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.015 [2024-07-15 18:12:49.514627] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.015 qpair failed and we were unable to recover it. 00:26:56.015 [2024-07-15 18:12:49.514856] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.514868] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.515095] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.515108] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.515364] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.515377] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.515580] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.515592] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.515787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.515799] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.516072] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.516084] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.516288] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.516300] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.516530] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.516543] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.516706] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.516717] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.516966] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.516977] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.517205] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.517217] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.517451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.517463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.517657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.517669] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.517921] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.517932] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.518161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.518173] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.518440] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.518452] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.518706] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.518718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.518844] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.518857] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.518985] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.518996] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.519175] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.519187] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.519453] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.519465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.519711] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.519723] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.519893] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.519906] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.520073] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.520085] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.520336] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.520348] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.520539] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.520550] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.520751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.520763] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.016 [2024-07-15 18:12:49.520874] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.016 [2024-07-15 18:12:49.520885] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.016 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.521137] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.521149] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.521322] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.521334] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.521526] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.521538] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.521714] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.521727] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.521977] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.521989] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.522240] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.522252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.522491] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.522503] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.522698] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.522711] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.522938] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.522949] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.523145] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.523157] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.523427] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.523439] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.523615] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.523627] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.523877] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.523889] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.524063] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.524075] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.524190] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.524202] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.524392] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.524404] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.524596] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.524608] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.524780] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.524792] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.525020] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.525032] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.525278] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.525292] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.525391] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.525403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.525664] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.525675] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.525936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.525948] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.526183] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.526196] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.526422] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.526434] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.526611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.526623] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.526891] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.017 [2024-07-15 18:12:49.526903] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.017 qpair failed and we were unable to recover it. 00:26:56.017 [2024-07-15 18:12:49.527130] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.527143] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.527330] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.527344] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.527443] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.527456] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.527705] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.527719] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.527901] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.527914] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.528020] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.528033] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.528283] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.528296] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.528537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.528551] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.528803] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.528817] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.529041] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.529054] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.529256] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.529268] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.529428] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.529440] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.529693] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.529705] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.529868] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.529881] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.530107] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.530121] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.530283] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.530296] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.530472] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.530485] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.530646] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.530658] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.530761] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.530774] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.531063] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.531075] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.531307] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.531321] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.531447] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.531460] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.531687] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.531699] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.531857] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.531870] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.532139] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.532151] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.532377] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.532390] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.532560] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.532573] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.532802] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.018 [2024-07-15 18:12:49.532814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.018 qpair failed and we were unable to recover it. 00:26:56.018 [2024-07-15 18:12:49.533007] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.533020] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.533243] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.533256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.533424] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.533436] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.533662] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.533675] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.533927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.533941] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.534200] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.534213] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.534499] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.534513] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.534745] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.534758] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.534991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.535004] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.535241] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.535254] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.535508] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.535521] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.535708] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.535720] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.535902] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.535914] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.536171] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.536184] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.536363] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.536376] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.536603] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.536616] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.536789] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.536802] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.536923] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.536936] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.537119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.537131] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.537381] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.537393] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.537578] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.537590] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.537818] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.537829] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.538019] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.538030] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.538209] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.538221] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.538325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.538337] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.538499] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.538511] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.538749] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.538761] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.019 [2024-07-15 18:12:49.539008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.019 [2024-07-15 18:12:49.539021] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.019 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.539192] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.539204] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.539460] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.539472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.539696] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.539707] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.539974] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.539986] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.540155] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.540167] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.540352] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.540364] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.540620] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.540632] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.540885] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.540897] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.541154] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.541166] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.541404] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.541416] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.541641] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.541653] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.541923] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.541935] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.542112] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.542123] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.542289] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.542301] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.542524] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.542536] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.542640] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.542652] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.542839] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.542851] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.542971] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.542982] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.543233] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.543245] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.543497] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.543509] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.543773] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.543785] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.020 [2024-07-15 18:12:49.544014] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.020 [2024-07-15 18:12:49.544026] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.020 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.544249] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.544262] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.544536] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.544548] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.544745] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.544757] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.545012] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.545024] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.545273] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.545285] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.545529] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.545541] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.545795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.545806] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.545929] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.545941] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.546184] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.546196] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.546453] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.546466] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.546734] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.546746] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.546975] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.546986] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.547232] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.547244] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.547423] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.547436] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.547612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.547624] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.547879] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.547891] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.547986] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.547999] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.548229] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.548241] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.548405] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.548416] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.548651] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.548664] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.548944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.548956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.549205] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.549219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.549449] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.549461] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.549619] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.549630] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.549881] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.549892] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.550150] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.550161] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.021 [2024-07-15 18:12:49.550394] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.021 [2024-07-15 18:12:49.550405] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.021 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.550566] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.550578] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.550700] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.550711] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.550958] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.550970] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.551128] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.551141] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.551399] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.551411] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.551659] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.551670] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.551895] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.551906] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.552068] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.552080] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.552260] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.552282] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.552508] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.552519] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.552692] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.552703] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.552930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.552942] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.553146] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.553157] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.553409] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.553421] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.553694] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.553706] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.553880] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.553892] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.554145] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.554156] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.554319] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.554331] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.554491] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.554503] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.554622] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.554633] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.554754] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.554765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.554969] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.554981] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.555164] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.555176] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.555362] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.555374] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.555536] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.555548] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.555825] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.555836] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.556007] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.022 [2024-07-15 18:12:49.556019] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.022 qpair failed and we were unable to recover it. 00:26:56.022 [2024-07-15 18:12:49.556188] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.556200] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.556404] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.556415] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.556524] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.556535] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.556708] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.556720] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.556899] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.556910] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.557080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.557092] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.557348] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.557360] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.557547] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.557560] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.557739] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.557751] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.558007] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.558018] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.558181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.558193] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.558370] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.558382] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.558611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.558623] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.558851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.558863] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.559107] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.559119] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.559322] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.559334] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.559560] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.559572] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.559698] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.559709] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.559963] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.559975] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.560215] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.560241] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.560357] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.560369] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.560647] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.560659] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.560859] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.560870] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.561099] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.561110] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.561310] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.561322] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.561579] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.561590] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.561768] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.023 [2024-07-15 18:12:49.561779] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.023 qpair failed and we were unable to recover it. 00:26:56.023 [2024-07-15 18:12:49.562028] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.562040] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.562292] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.562304] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.562472] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.562484] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.562709] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.562722] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.562900] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.562911] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.563072] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.563084] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.563334] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.563346] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.563525] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.563537] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.563786] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.563797] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.564016] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.564028] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.564137] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.564149] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.564434] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.564446] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.564676] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.564688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.564856] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.564868] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.565037] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.565049] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.565222] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.565236] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.565463] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.565475] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.565638] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.565649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.565745] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.565757] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.565872] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.565884] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.566119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.566135] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.566242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.566255] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.566447] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.566461] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.566714] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.566725] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.566977] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.566989] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.567233] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.567245] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.567419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.567431] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.567679] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.024 [2024-07-15 18:12:49.567691] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.024 qpair failed and we were unable to recover it. 00:26:56.024 [2024-07-15 18:12:49.567884] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.567896] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.568148] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.568161] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.568398] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.568411] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.568600] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.568613] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.568848] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.568861] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.569057] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.569070] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.569278] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.569291] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.569540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.569552] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.569724] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.569736] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.569872] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:26:56.025 [2024-07-15 18:12:49.569903] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:26:56.025 [2024-07-15 18:12:49.569911] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:26:56.025 [2024-07-15 18:12:49.569918] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:26:56.025 [2024-07-15 18:12:49.569924] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:26:56.025 [2024-07-15 18:12:49.569963] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.569975] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.570034] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 5 00:26:56.025 [2024-07-15 18:12:49.570198] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.570209] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.570140] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 6 00:26:56.025 [2024-07-15 18:12:49.570261] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 4 00:26:56.025 [2024-07-15 18:12:49.570319] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.570330] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.570261] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 7 00:26:56.025 [2024-07-15 18:12:49.570533] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.570545] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.570746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.570757] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.570995] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.571007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.571236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.571248] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.571513] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.571525] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.571630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.571642] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.571867] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.571879] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.571992] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.572004] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.572179] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.572191] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.572360] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.572372] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.572592] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.572603] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.572777] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.025 [2024-07-15 18:12:49.572788] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.025 qpair failed and we were unable to recover it. 00:26:56.025 [2024-07-15 18:12:49.573011] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.573023] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.573248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.573260] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.573388] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.573399] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.573640] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.573651] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.573879] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.573891] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.574175] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.574188] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.574364] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.574377] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.574616] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.574628] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.574884] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.574896] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.575073] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.575085] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.575246] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.575259] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.575486] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.575499] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.575665] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.575677] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.575912] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.575924] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.576176] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.576189] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.576446] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.576458] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.576689] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.576700] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.576866] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.576878] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.577044] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.577058] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.577248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.577260] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.577485] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.577498] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.577773] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.577786] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.578030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.578042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.578272] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.578284] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.578533] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.578545] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.578674] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.026 [2024-07-15 18:12:49.578686] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.026 qpair failed and we were unable to recover it. 00:26:56.026 [2024-07-15 18:12:49.578790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.578802] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.579054] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.579067] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.579255] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.579268] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.579523] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.579535] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.579764] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.579777] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.579938] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.579950] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.580186] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.580199] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.580299] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.580311] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.580546] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.580558] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.580832] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.580844] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.581077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.581090] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.581321] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.581333] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.581606] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.581619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.581800] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.581813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.582012] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.582024] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.582145] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.582157] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.582266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.582279] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.582383] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.582395] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.582574] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.582587] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.582841] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.582884] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.583178] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.583215] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.583516] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.583553] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.583747] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.027 [2024-07-15 18:12:49.583765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.027 qpair failed and we were unable to recover it. 00:26:56.027 [2024-07-15 18:12:49.584026] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.584041] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.584282] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.584299] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.584479] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.584495] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.584751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.584766] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.585026] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.585041] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.585145] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.585160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.585428] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.585445] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.585684] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.585699] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.585881] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.585895] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.586131] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.586146] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.586389] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.586405] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.586613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.586628] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.586790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.586806] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.587061] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.587077] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.587334] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.587351] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.587543] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.587558] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.587769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.587785] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.588043] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.588058] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.588323] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.588340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.588550] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.588565] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.588701] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.588716] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.588897] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.588913] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.589092] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.589108] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.589294] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.589315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.589577] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.589593] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.589775] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.589790] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.590037] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.028 [2024-07-15 18:12:49.590052] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.028 qpair failed and we were unable to recover it. 00:26:56.028 [2024-07-15 18:12:49.590255] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.590271] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.590451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.590468] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.590723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.590739] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.590950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.590966] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.591154] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.591170] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.591352] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.591368] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.591551] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.591566] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.591825] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.591841] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.592092] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.592107] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.592360] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.592377] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.592558] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.592573] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.592750] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.592765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.593027] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.593042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.593221] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.593242] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.593439] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.593454] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.593644] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.593659] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.593840] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.593855] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.594099] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.594115] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.594301] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.594317] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.594575] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.594591] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.594767] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.594783] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.595044] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.595059] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.595312] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.595328] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.595569] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.595586] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.595756] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.595771] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.595956] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.595972] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.596233] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.596250] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.596437] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.596453] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.596653] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.029 [2024-07-15 18:12:49.596669] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.029 qpair failed and we were unable to recover it. 00:26:56.029 [2024-07-15 18:12:49.596793] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.596809] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.596914] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.596930] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.597058] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.597075] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.597282] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.597298] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.597559] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.597576] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.597854] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.597871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.598110] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.598127] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.598386] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.598404] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.598536] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.598553] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.598763] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.598781] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.598987] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.599005] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.599272] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.599290] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.599470] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.599486] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.599774] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.599792] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.600009] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.600029] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.600273] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.600292] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.600559] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.600578] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.600841] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.600859] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.601096] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.601114] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.601329] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.601348] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.601482] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.601499] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.601752] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.601772] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.601974] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.601992] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.602106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.602122] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.602304] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.602322] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.602451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.602469] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.602587] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.602603] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.602819] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.602835] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.603071] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.603088] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.603344] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.030 [2024-07-15 18:12:49.603362] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.030 qpair failed and we were unable to recover it. 00:26:56.030 [2024-07-15 18:12:49.603613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.603630] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.603799] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.603815] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.604079] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.604096] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.604269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.604286] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.604537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.604554] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.604815] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.604833] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.605081] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.605097] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.605356] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.605373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.605613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.605630] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.605865] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.605882] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.606057] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.606075] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.606268] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.606285] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.606521] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.606539] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.606836] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.606852] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.607115] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.607131] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.607301] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.607317] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.607447] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.607463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.607654] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.607669] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.607878] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.607894] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.608093] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.608108] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.608388] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.608404] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.608666] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.608681] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.608893] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.608908] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.609175] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.609191] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.609451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.609467] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.609679] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.609695] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.609958] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.609974] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.610234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.610250] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.610483] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.610500] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.031 [2024-07-15 18:12:49.610768] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.031 [2024-07-15 18:12:49.610784] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.031 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.611027] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.611044] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.611272] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.611291] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.611501] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.611535] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.611740] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.611757] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.612037] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.612053] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.612183] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.612199] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.612436] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.612451] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.612722] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.612737] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.612903] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.612918] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.613233] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.613249] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.613510] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.613526] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.613717] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.613732] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.614015] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.614031] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.614265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.614281] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.614477] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.614494] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.614727] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.614748] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.615024] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.615039] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.615298] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.615314] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.615490] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.615506] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.615692] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.615708] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.615880] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.615897] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.616071] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.616089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.616355] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.616371] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.616572] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.616589] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.616793] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.616812] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.617077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.617097] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.617340] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.617358] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.617620] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.617638] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.617827] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.032 [2024-07-15 18:12:49.617844] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.032 qpair failed and we were unable to recover it. 00:26:56.032 [2024-07-15 18:12:49.618039] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.618056] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.618316] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.618336] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.618508] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.618527] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.618801] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.618820] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.619080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.619097] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.619335] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.619351] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.619613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.619629] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.619740] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.619755] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.619932] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.619947] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.620187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.620202] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.620417] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.620433] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.620708] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.620725] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.620993] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.621008] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.621251] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.621280] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.621556] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.621572] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.621775] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.621790] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.622023] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.622038] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.622343] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.622359] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.622619] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.622635] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.622804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.622820] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.623056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.623071] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.623278] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.623294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.623492] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.623507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.623790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.033 [2024-07-15 18:12:49.623806] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.033 qpair failed and we were unable to recover it. 00:26:56.033 [2024-07-15 18:12:49.623986] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.624002] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.624177] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.624192] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.624313] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.624328] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.624505] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.624520] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.624691] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.624706] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.624946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.624962] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.625150] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.625166] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.625403] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.625419] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.625678] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.625695] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.625943] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.625960] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.626150] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.626166] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.626342] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.626358] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.626530] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.626545] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.626806] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.626821] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.627009] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.627025] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.627272] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.627289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.627550] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.627568] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.627836] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.627851] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.628059] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.628074] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.628263] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.628278] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.628532] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.628547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.628821] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.628836] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.629120] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.629135] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.629377] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.629392] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.629598] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.629613] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.629807] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.629822] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.630038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.630053] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.630256] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.630273] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.630535] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.630550] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.630784] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.034 [2024-07-15 18:12:49.630800] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.034 qpair failed and we were unable to recover it. 00:26:56.034 [2024-07-15 18:12:49.630980] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.630995] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.631256] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.631272] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.631476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.631492] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.631685] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.631700] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.631938] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.631955] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.632193] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.632209] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.632405] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.632420] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.632598] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.632613] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.632850] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.632865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.633051] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.633066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.633244] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.633259] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.633388] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.633403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.633573] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.633589] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.633780] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.633797] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.634058] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.634075] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.634341] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.634360] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.634575] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.634594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.634763] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.634779] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.635014] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.635029] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.635291] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.635307] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.635490] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.635506] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.635800] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.635815] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.636001] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.636016] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.636191] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.636207] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.636336] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.636351] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.636537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.636553] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.636728] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.636744] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.636990] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.637006] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.637206] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.637221] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.637513] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.637528] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.637733] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.637748] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.637934] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.637950] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.638062] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.638078] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.638207] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.638222] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.638345] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.638360] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.638584] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.035 [2024-07-15 18:12:49.638600] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.035 qpair failed and we were unable to recover it. 00:26:56.035 [2024-07-15 18:12:49.638834] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.638850] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.639097] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.639113] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.639373] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.639390] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.639639] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.639655] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.639855] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.639870] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.640110] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.640127] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.640372] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.640388] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.640580] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.640596] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.640880] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.640896] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.641150] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.641165] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.641451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.641467] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.641747] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.641764] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.642047] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.642062] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.642245] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.642261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.642518] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.642533] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.642802] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.642818] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.643031] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.643046] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.643258] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.643275] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.643567] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.643599] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.643876] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.643892] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.644081] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.644096] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.644382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.644401] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.644589] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.644605] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.644772] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.644787] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.644971] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.644986] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.645167] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.645183] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.645438] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.645454] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.645573] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.645588] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.645809] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.645825] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.646013] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.646029] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.646290] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.646306] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.646592] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.646613] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.646858] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.646874] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.647060] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.647076] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.647205] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.647221] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.647409] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.647425] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.647628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.647643] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.647901] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.647916] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.648096] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.648111] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.648240] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.648255] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.648515] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.648531] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.648711] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.648726] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.648929] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.648945] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.649153] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.649168] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.649371] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.649386] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.649593] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.649608] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.036 [2024-07-15 18:12:49.649844] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.036 [2024-07-15 18:12:49.649860] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.036 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.650111] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.650126] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.650364] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.650379] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.650558] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.650573] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.650830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.650845] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.651033] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.651049] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.651158] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.651173] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.651396] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.651411] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.651607] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.651622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.651857] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.651872] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.652051] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.652066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.652303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.652319] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.652461] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.652479] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.652608] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.652623] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.652794] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.652809] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.653010] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.653025] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.653134] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.653149] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.653405] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.653421] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.653664] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.653679] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.653919] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.653934] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.654193] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.654208] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.654383] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.654399] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.654583] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.654599] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.654844] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.654859] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.655037] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.655053] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.655288] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.655304] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.655570] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.655585] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.655830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.655846] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.656100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.656116] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.656384] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.656400] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.656681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.656696] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.656867] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.656882] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.657072] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.657087] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.657347] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.657363] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.657610] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.657626] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.657879] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.657895] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.037 qpair failed and we were unable to recover it. 00:26:56.037 [2024-07-15 18:12:49.658077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.037 [2024-07-15 18:12:49.658092] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.658299] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.658315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.658489] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.658505] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.658687] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.658706] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.658944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.658959] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.659129] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.659144] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.659408] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.659423] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.659685] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.659701] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.659901] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.659916] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.660171] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.660187] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.660409] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.660425] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.660660] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.660676] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.660959] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.660975] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.661162] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.661177] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.661365] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.661380] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.661617] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.661632] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.661841] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.661856] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.662113] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.662128] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.662371] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.662387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.662556] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.662570] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.662828] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.662844] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.663030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.663046] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.663245] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.663261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.663477] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.663493] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.663728] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.663743] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.663935] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.663950] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.664128] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.664144] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.664448] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.664464] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.664650] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.664666] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.664900] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.664915] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.665185] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.665202] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.665429] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.665444] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.665611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.665627] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.665757] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.665772] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.665940] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.665956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.666141] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.666156] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.666406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.666422] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.666604] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.666620] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.666805] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.666820] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.667004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.667020] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.667253] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.667268] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.667527] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.667543] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.667728] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.667744] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.667912] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.667927] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.668125] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.668141] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.668387] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.668402] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.668634] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.668649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.038 qpair failed and we were unable to recover it. 00:26:56.038 [2024-07-15 18:12:49.668908] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.038 [2024-07-15 18:12:49.668923] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.669091] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.669106] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.669223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.669241] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.669488] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.669504] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.669682] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.669698] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.669802] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.669817] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.669994] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.670010] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.670249] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.670265] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.670435] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.670450] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.670591] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.670607] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.670731] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.670749] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.670963] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.670979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.671164] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.671179] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.671444] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.671459] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.671717] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.671732] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.671903] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.671918] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.672089] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.672104] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.672291] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.672307] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.672511] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.672527] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.672772] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.672787] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.673029] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.673045] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.673281] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.673297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.673558] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.673574] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.673821] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.673836] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.674164] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.674195] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.674478] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.674498] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.674759] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.674775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.675008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.675023] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.675307] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.675323] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.675581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.675597] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.675781] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.675797] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.676092] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.676107] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.676341] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.676357] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.676640] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.676656] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.676910] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.676925] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.677055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.677070] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.677306] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.677323] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.677515] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.677533] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.677790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.677806] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.677937] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.677952] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.678189] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.678205] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.678464] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.678480] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.678692] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.678706] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.039 qpair failed and we were unable to recover it. 00:26:56.039 [2024-07-15 18:12:49.678964] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.039 [2024-07-15 18:12:49.678979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.679261] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.679277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.679537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.679552] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.679833] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.679848] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.680105] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.680120] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.680384] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.680400] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.680665] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.680680] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.680885] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.680900] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.681071] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.681087] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.681324] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.681341] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.681545] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.681560] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.681801] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.681817] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.681988] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.682004] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.682254] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.682270] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.682457] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.682472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.682706] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.682721] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.682902] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.682919] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.683100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.683115] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.683366] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.683382] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.683686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.683701] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.683939] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.683954] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.684147] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.684166] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.684376] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.684392] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.684567] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.684578] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.684818] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.684830] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.685015] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.685028] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.685195] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.685207] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.685385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.685397] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.685650] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.685662] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.685916] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.685928] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.686106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.686118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.686233] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.686245] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.686419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.686431] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.686686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.686698] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.686951] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.686968] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.687202] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.687214] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.687393] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.687405] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.687581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.687594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.687765] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.687777] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.688003] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.688015] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.688187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.688199] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.688443] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.688456] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.688681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.688692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.688798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.688810] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.689057] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.689068] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.689303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.689314] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.689511] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.689522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.040 [2024-07-15 18:12:49.689747] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.040 [2024-07-15 18:12:49.689759] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.040 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.690011] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.690023] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.690229] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.690241] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.690512] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.690523] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.690703] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.690715] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.690943] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.690956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.691185] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.691197] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.691389] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.691402] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.691629] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.691640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.691803] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.691816] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.692044] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.692056] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.692252] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.692264] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.692519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.692530] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.692766] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.692777] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.692972] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.692984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.693236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.693248] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.693494] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.693506] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.693686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.693698] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.693922] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.693933] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.694159] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.694170] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.694396] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.694408] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.694578] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.694590] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.694688] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.694701] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.694897] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.694908] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.695160] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.695172] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.695433] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.695446] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.695676] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.695688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.695915] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.695929] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.696174] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.696187] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.696383] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.696395] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.696590] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.696602] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.696833] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.696844] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.697017] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.697028] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.697144] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.697156] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.697327] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.697339] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.697593] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.697605] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.697831] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.697842] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.698043] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.698055] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.698180] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.698192] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.698371] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.698384] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.698620] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.698631] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.041 [2024-07-15 18:12:49.698899] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.041 [2024-07-15 18:12:49.698911] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.041 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.699173] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.699185] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.699303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.699315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.699540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.699551] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.699777] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.699789] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.699980] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.699991] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.700238] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.700250] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.700411] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.700423] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.700673] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.700684] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.700897] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.700909] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.701175] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.701187] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.701441] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.701452] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.701578] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.701590] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.701699] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.701711] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.701962] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.701974] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.702144] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.702156] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.702331] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.702343] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.702542] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.702554] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.702723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.702734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.702988] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.702999] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.703253] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.703265] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.042 [2024-07-15 18:12:49.703506] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.042 [2024-07-15 18:12:49.703517] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.042 qpair failed and we were unable to recover it. 00:26:56.327 [2024-07-15 18:12:49.703691] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.327 [2024-07-15 18:12:49.703704] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.327 qpair failed and we were unable to recover it. 00:26:56.327 [2024-07-15 18:12:49.703957] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.327 [2024-07-15 18:12:49.703971] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.327 qpair failed and we were unable to recover it. 00:26:56.327 [2024-07-15 18:12:49.704087] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.327 [2024-07-15 18:12:49.704099] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.327 qpair failed and we were unable to recover it. 00:26:56.327 [2024-07-15 18:12:49.704277] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.327 [2024-07-15 18:12:49.704290] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.327 qpair failed and we were unable to recover it. 00:26:56.327 [2024-07-15 18:12:49.704473] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.327 [2024-07-15 18:12:49.704487] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.327 qpair failed and we were unable to recover it. 00:26:56.327 [2024-07-15 18:12:49.704595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.327 [2024-07-15 18:12:49.704606] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.327 qpair failed and we were unable to recover it. 00:26:56.327 [2024-07-15 18:12:49.704775] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.327 [2024-07-15 18:12:49.704788] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.327 qpair failed and we were unable to recover it. 00:26:56.327 [2024-07-15 18:12:49.704964] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.327 [2024-07-15 18:12:49.704975] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.327 qpair failed and we were unable to recover it. 00:26:56.327 [2024-07-15 18:12:49.705172] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.327 [2024-07-15 18:12:49.705183] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.327 qpair failed and we were unable to recover it. 00:26:56.327 [2024-07-15 18:12:49.705345] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.327 [2024-07-15 18:12:49.705357] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.327 qpair failed and we were unable to recover it. 00:26:56.327 [2024-07-15 18:12:49.705529] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.327 [2024-07-15 18:12:49.705540] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.327 qpair failed and we were unable to recover it. 00:26:56.327 [2024-07-15 18:12:49.705791] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.705803] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.706079] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.706091] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.706221] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.706236] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.706487] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.706499] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.706605] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.706617] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.706808] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.706819] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.707059] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.707071] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.707303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.707316] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.707512] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.707525] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.707695] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.707706] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.707932] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.707943] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.708169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.708182] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.708412] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.708424] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.708676] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.708687] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.708850] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.708862] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.709054] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.709066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.709335] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.709348] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.709525] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.709537] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.709797] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.709809] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.710062] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.710074] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.710236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.710250] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.710429] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.710441] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.710620] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.710632] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.710796] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.710808] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.710968] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.710980] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.711208] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.711220] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.711334] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.711346] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.711614] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.711625] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.711818] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.711830] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.712002] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.712014] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.712269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.712281] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.712527] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.712539] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.712777] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.712789] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.712954] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.712966] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.713221] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.713235] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.713405] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.713417] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.713664] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.713676] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.713929] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.713941] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.714182] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.714194] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.714366] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.328 [2024-07-15 18:12:49.714378] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.328 qpair failed and we were unable to recover it. 00:26:56.328 [2024-07-15 18:12:49.714541] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.714554] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.714815] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.714828] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.714995] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.715007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.715260] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.715273] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.715520] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.715532] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.715773] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.715784] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.716008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.716020] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.716286] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.716298] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.716496] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.716508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.716797] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.716808] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.716978] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.716990] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.717240] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.717252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.717498] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.717510] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.717608] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.717620] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.717855] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.717866] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.718100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.718111] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.718297] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.718309] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.718482] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.718494] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.718656] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.718668] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.718893] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.718905] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.719160] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.719175] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.719346] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.719358] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.719524] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.719537] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.719752] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.719764] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.719935] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.719946] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.720179] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.720191] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.720359] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.720371] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.720611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.720623] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.720782] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.720793] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.721038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.721050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.721222] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.721238] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.721490] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.721501] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.721687] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.721700] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.721950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.721962] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.722075] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.722087] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.722339] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.722350] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.722577] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.722588] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.722847] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.722859] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.722976] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.722987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.723161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.723173] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.329 [2024-07-15 18:12:49.723416] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.329 [2024-07-15 18:12:49.723429] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.329 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.723654] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.723666] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.723942] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.723954] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.724115] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.724127] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.724299] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.724310] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.724477] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.724488] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.724760] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.724771] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.724980] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.724992] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.725260] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.725272] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.725383] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.725394] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.725637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.725648] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.725873] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.725884] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.726052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.726064] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.726300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.726312] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.726420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.726432] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.726660] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.726671] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.726899] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.726911] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.727087] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.727099] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.727324] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.727335] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.727530] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.727542] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.727792] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.727806] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.728048] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.728060] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.728311] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.728323] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.728495] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.728506] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.728761] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.728773] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.728933] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.728944] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.729136] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.729147] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.729337] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.729349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.729612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.729623] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.729799] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.729811] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.729996] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.730007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.730201] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.730213] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.730397] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.730409] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.730590] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.730602] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.730831] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.730842] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.731024] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.731036] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.731265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.731277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.731384] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.731396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.731506] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.731517] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.731706] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.731719] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.330 [2024-07-15 18:12:49.731945] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.330 [2024-07-15 18:12:49.731957] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.330 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.732207] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.732219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.732385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.732397] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.732616] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.732627] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.732864] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.732876] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.733127] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.733140] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.733314] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.733326] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.733443] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.733455] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.733634] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.733645] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.733822] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.733833] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.734015] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.734026] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.734227] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.734239] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.734426] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.734438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.734692] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.734704] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.734931] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.734943] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.735122] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.735134] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.735337] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.735349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.735530] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.735541] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.735713] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.735724] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.735979] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.735991] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.736160] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.736174] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.736348] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.736361] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.736543] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.736555] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.736778] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.736790] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.736993] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.737005] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.737240] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.737253] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.737444] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.737456] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.737708] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.737720] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.737884] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.737895] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.738147] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.738160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.738415] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.738426] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.738601] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.738612] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.738783] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.738794] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.739043] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.739054] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.739303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.739315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.739424] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.739436] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.739548] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.739560] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.331 [2024-07-15 18:12:49.739721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.331 [2024-07-15 18:12:49.739733] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.331 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.739968] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.739980] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.740208] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.740219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.740496] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.740507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.740617] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.740629] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.740879] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.740890] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.741156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.741167] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.741414] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.741427] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.741602] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.741615] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.741773] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.741784] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.742032] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.742044] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.742157] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.742169] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.742433] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.742445] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.742606] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.742618] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.742815] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.742827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.743081] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.743094] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.743208] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.743220] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.743475] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.743487] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.743671] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.743682] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.743928] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.743940] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.744191] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.744203] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.744382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.744394] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.744557] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.744570] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.744822] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.744835] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.745103] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.745115] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.745318] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.745330] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.745535] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.745547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.745785] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.745797] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.745981] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.745993] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.746169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.746181] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.746292] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.746303] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.746551] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.746562] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.746792] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.746803] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.746965] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.746977] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.747198] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.747210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.747431] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.747443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.747676] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.747688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.747851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.747863] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.332 qpair failed and we were unable to recover it. 00:26:56.332 [2024-07-15 18:12:49.748119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.332 [2024-07-15 18:12:49.748131] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.748401] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.748413] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.748641] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.748653] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.748816] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.748827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.749074] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.749086] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.749284] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.749296] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.749492] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.749504] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.749777] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.749789] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.749972] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.749984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.750167] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.750179] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.750345] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.750358] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.750521] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.750533] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.750696] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.750707] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.750830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.750843] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.750943] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.750955] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.751155] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.751166] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.751375] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.751386] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.751557] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.751570] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.751810] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.751822] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.752019] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.752032] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.752204] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.752217] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.752334] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.752346] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.752588] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.752599] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.752780] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.752791] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.753029] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.753040] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.753293] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.753307] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.753434] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.753446] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.753617] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.753629] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.753809] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.753821] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.754074] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.754087] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.754313] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.754324] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.754554] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.754566] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.754756] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.754768] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.755018] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.755029] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.755211] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.755223] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.755401] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.755413] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.755641] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.755652] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.755927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.755938] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.333 [2024-07-15 18:12:49.756163] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.333 [2024-07-15 18:12:49.756175] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.333 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.756423] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.756435] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.756611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.756623] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.756872] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.756883] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.757095] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.757107] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.757281] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.757293] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.757531] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.757543] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.757813] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.757826] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.757947] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.757959] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.758131] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.758143] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.758369] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.758381] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.758558] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.758569] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.758835] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.758846] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.759094] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.759106] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.759350] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.759362] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.759632] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.759643] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.759899] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.759911] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.760160] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.760171] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.760348] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.760372] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.760624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.760636] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.760745] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.760756] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.760917] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.760930] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.761156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.761168] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.761340] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.761352] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.761515] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.761527] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.761754] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.761765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.761938] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.761950] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.762179] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.762193] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.762420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.762432] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.762599] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.762611] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.762864] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.762876] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.763038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.763049] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.763301] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.763313] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.763440] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.763451] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.763679] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.763690] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.763892] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.763904] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.764007] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.764019] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.764180] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.764192] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.764356] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.764368] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.764595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.334 [2024-07-15 18:12:49.764607] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.334 qpair failed and we were unable to recover it. 00:26:56.334 [2024-07-15 18:12:49.764776] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.764787] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.764890] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.764902] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.765084] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.765096] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.765358] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.765370] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.765596] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.765608] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.765792] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.765803] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.766050] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.766062] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.766222] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.766244] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.766495] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.766507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.766756] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.766768] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.767012] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.767023] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.767222] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.767237] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.767461] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.767473] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.767636] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.767648] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.767852] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.767864] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.768116] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.768127] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.768403] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.768415] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.768586] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.768598] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.768824] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.768836] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.769005] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.769017] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.769260] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.769272] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.769472] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.769484] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.769686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.769697] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.769862] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.769874] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.769982] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.769994] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.770242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.770253] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.770379] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.770390] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.770636] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.770649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.770879] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.770891] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.771052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.771063] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.771277] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.771289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.771465] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.771477] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.771717] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.771728] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.771958] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.771970] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.772197] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.772209] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.772410] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.772423] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.772583] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.335 [2024-07-15 18:12:49.772595] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.335 qpair failed and we were unable to recover it. 00:26:56.335 [2024-07-15 18:12:49.772791] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.772803] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.772982] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.772994] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.773242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.773254] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.773510] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.773521] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.773706] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.773717] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.773945] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.773957] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.774199] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.774210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.774471] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.774483] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.774590] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.774602] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.774796] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.774808] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.775038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.775050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.775167] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.775178] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.775373] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.775385] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.775483] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.775495] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.775679] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.775691] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.775953] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.775964] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.776141] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.776153] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.776406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.776418] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.776609] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.776620] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.776903] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.776914] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.777095] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.336 [2024-07-15 18:12:49.777107] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.336 qpair failed and we were unable to recover it. 00:26:56.336 [2024-07-15 18:12:49.777298] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.777310] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.777484] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.777496] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.777664] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.777676] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.777902] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.777914] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.778159] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.778170] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.778344] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.778356] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.778528] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.778539] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.778737] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.778750] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.779001] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.779013] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.779201] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.779216] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.779316] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.779328] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.779578] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.779589] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.779819] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.779831] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.779997] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.780009] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.780254] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.780265] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.780519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.780531] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.780777] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.780790] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.781067] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.781079] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.781314] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.781326] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.781553] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.781564] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.781807] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.781819] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.782080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.782092] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.782344] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.782356] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.782538] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.782549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.782708] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.782719] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.782840] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.782851] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.783119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.783131] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.783330] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.783341] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.783616] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.783628] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.783786] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.783797] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.784026] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.784038] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.784169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.784181] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.784355] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.784367] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.784617] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.784629] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.337 [2024-07-15 18:12:49.784879] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.337 [2024-07-15 18:12:49.784891] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.337 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.785139] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.785151] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.785332] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.785345] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.785537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.785549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.785660] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.785672] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.785853] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.785865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.786025] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.786037] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.786229] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.786241] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.786465] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.786476] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.786702] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.786714] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.786914] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.786926] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.787087] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.787100] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.787345] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.787357] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.787539] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.787550] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.787753] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.787765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.787946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.787960] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.788209] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.788220] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.788412] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.788423] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.788650] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.788662] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.788919] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.788931] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.789125] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.789137] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.789361] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.789373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.789629] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.789641] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.789895] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.789907] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.790139] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.790151] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.790405] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.790417] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.790585] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.790597] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.790721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.790732] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.790976] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.790987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.791174] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.791185] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.791412] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.791424] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.791682] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.791694] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.791944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.791956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.792201] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.792213] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.792463] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.792474] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.792700] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.792712] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.792943] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.792955] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.793116] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.793128] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.793286] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.793298] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.793557] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.793569] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.793747] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.793758] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.338 [2024-07-15 18:12:49.794008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.338 [2024-07-15 18:12:49.794020] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.338 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.794213] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.794228] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.794501] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.794513] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.794717] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.794728] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.794997] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.795008] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.795269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.795281] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.795446] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.795459] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.795703] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.795714] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.795820] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.795831] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.796077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.796088] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.796292] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.796304] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.796468] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.796480] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.796733] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.796744] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.796993] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.797004] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.797245] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.797259] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.797530] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.797542] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.797776] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.797789] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.798033] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.798045] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.798273] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.798285] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.798393] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.798405] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.798601] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.798613] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.798838] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.798849] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.799076] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.799088] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.799268] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.799283] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.799447] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.799458] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.799639] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.799652] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.799819] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.799831] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.800004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.800017] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.800304] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.800317] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.800455] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.800467] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.800696] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.800708] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.800871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.800884] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.801130] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.801143] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.801345] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.801357] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.801607] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.801619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.801813] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.801826] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.802077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.802089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.802335] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.339 [2024-07-15 18:12:49.802346] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.339 qpair failed and we were unable to recover it. 00:26:56.339 [2024-07-15 18:12:49.802526] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.802537] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.802706] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.802718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.802945] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.802958] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.803058] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.803070] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.803263] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.803275] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.803396] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.803407] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.803592] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.803604] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.803875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.803887] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.803996] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.804008] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.804197] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.804209] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.804454] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.804465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.804654] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.804666] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.804891] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.804903] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.805154] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.805166] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.805398] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.805411] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.805667] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.805679] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.805945] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.805960] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.806142] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.806155] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.806338] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.806351] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.806531] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.806543] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.806784] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.806797] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.807028] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.807040] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.807281] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.807293] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.807459] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.807471] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.807723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.807735] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.807897] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.807908] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.808158] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.808170] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.808298] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.808311] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.808556] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.808568] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.808774] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.808786] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.808975] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.808987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.809150] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.809162] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.809346] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.809358] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.809615] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.809627] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.809851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.809863] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.810089] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.810102] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.810349] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.810362] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.810494] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.810506] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.810613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.810625] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.810851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.340 [2024-07-15 18:12:49.810863] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.340 qpair failed and we were unable to recover it. 00:26:56.340 [2024-07-15 18:12:49.811114] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.811126] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.811354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.811365] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.811530] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.811543] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.811788] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.811801] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.812004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.812016] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.812187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.812200] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.812397] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.812409] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.812603] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.812615] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.812795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.812809] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.812914] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.812926] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.813153] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.813164] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.813350] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.813362] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.813601] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.813613] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.813872] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.813884] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.814110] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.814122] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.814375] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.814388] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.814618] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.814632] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.814892] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.814903] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.815074] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.815086] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.815282] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.815295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.815479] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.815491] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.815660] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.815672] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.815871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.815884] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.816060] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.816072] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.816325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.816338] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.816604] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.816616] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.816843] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.816854] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.817040] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.817052] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.817310] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.817322] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.817433] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.817444] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.817720] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.341 [2024-07-15 18:12:49.817732] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.341 qpair failed and we were unable to recover it. 00:26:56.341 [2024-07-15 18:12:49.817905] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.817916] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.818165] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.818176] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.818371] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.818384] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.818611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.818622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.818752] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.818763] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.818957] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.818969] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.819194] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.819206] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.819384] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.819396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.819613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.819625] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.819788] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.819800] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.819986] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.819998] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.820095] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.820107] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.820266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.820280] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.820376] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.820388] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.820559] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.820570] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.820747] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.820760] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.821033] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.821045] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.821206] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.821219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.821475] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.821487] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.821692] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.821704] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.821801] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.821813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.822045] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.822057] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.822237] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.822249] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.822495] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.822507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.822667] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.822679] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.822843] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.822855] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.823052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.823064] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.823310] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.823322] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.823548] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.823561] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.823730] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.823742] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.823913] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.823924] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.824117] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.824129] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.824293] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.824306] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.824480] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.824492] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.824720] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.824732] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.824983] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.824995] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.825174] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.825186] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.825291] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.825303] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.825494] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.825505] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.825738] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.342 [2024-07-15 18:12:49.825751] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.342 qpair failed and we were unable to recover it. 00:26:56.342 [2024-07-15 18:12:49.826003] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.826015] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.826196] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.826208] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.826413] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.826425] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.826611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.826623] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.826785] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.826797] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.826968] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.826980] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.827205] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.827217] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.827441] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.827453] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.827605] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.827618] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.827862] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.827873] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.827985] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.827997] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.828239] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.828252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.828353] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.828368] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.828472] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.828483] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.828709] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.828720] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.828895] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.828907] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.829066] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.829078] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.829248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.829260] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.829453] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.829465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.829627] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.829639] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.829800] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.829813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.829996] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.830008] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.830184] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.830197] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.830275] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.830287] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.830498] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.830510] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.830641] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.830653] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.830830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.830842] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.830963] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.830975] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.831087] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.831100] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.831232] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.831244] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.831415] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.831427] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.831600] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.831612] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.831787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.831799] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.832036] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.832048] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.832274] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.832287] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.832447] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.832459] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.343 qpair failed and we were unable to recover it. 00:26:56.343 [2024-07-15 18:12:49.832634] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.343 [2024-07-15 18:12:49.832647] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.832762] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.832774] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.832886] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.832898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.833062] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.833074] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.833300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.833312] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.833481] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.833493] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.833613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.833625] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.833812] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.833824] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.833931] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.833943] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.834101] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.834113] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.834240] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.834253] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.834428] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.834440] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.834639] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.834651] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.834765] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.834777] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.834957] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.834969] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.835134] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.835146] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.835326] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.835340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.835569] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.835581] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.835807] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.835819] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.835920] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.835932] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.836180] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.836192] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.836311] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.836323] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.836482] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.836495] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.836673] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.836684] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.836859] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.836871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.837122] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.837134] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.837236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.837248] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.837413] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.837425] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.837693] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.837705] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.837936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.837948] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.838180] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.838192] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.838383] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.838395] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.838592] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.838604] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.838706] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.838718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.838882] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.838894] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.839066] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.839078] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.839172] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.839185] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.839300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.839312] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.839539] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.839552] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.839778] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.839789] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.344 qpair failed and we were unable to recover it. 00:26:56.344 [2024-07-15 18:12:49.840016] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.344 [2024-07-15 18:12:49.840028] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.840192] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.840204] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.840389] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.840402] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.840644] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.840656] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.840820] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.840832] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.840999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.841011] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.841196] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.841209] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.841370] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.841382] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.841586] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.841598] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.841798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.841811] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.841999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.842011] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.842183] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.842195] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.842358] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.842370] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.842533] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.842546] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.842724] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.842735] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.842849] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.842861] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.843088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.843102] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.843229] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.843242] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.843436] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.843448] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.843698] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.843710] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.843877] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.843889] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.844048] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.844059] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.844288] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.844300] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.844527] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.844539] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.844647] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.844659] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.844753] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.844765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.844884] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.844895] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.844976] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.844987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.845090] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.845102] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.845257] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.845269] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.845442] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.845454] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.845625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.845637] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.845739] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.845751] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.845850] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.845862] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.846034] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.846045] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.345 qpair failed and we were unable to recover it. 00:26:56.345 [2024-07-15 18:12:49.846271] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.345 [2024-07-15 18:12:49.846283] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.846519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.846530] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.846637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.846648] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.846819] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.846831] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.847068] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.847080] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.847240] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.847253] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.847421] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.847433] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.847593] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.847605] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.847714] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.847726] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.847828] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.847839] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.847996] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.848008] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.848118] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.848130] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.848286] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.848298] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.848401] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.848412] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.848517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.848529] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.848762] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.848773] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.848999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.849010] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.849125] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.849137] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.849232] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.849243] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.849363] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.849375] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.849496] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.849508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.849616] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.849629] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.849886] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.849897] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.850084] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.850096] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.850334] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.850346] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.850451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.850463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.850687] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.850699] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.850930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.850942] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.851106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.851117] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.851233] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.851246] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.851448] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.851461] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.851638] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.851650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.851812] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.851824] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.851999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.852010] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.852128] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.852139] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.852261] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.852273] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.852442] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.852453] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.852652] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.852664] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.852827] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.852839] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.853081] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.346 [2024-07-15 18:12:49.853093] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.346 qpair failed and we were unable to recover it. 00:26:56.346 [2024-07-15 18:12:49.853197] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.853208] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.853330] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.853342] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.853439] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.853450] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.853567] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.853579] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.853741] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.853752] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.853931] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.853942] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.854048] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.854059] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.854245] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.854257] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.854489] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.854500] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.854660] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.854672] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.854843] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.854855] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.855019] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.855030] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.855190] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.855201] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.855366] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.855378] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.855549] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.855561] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.855745] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.855757] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.855860] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.855872] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.855992] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.856004] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.856205] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.856216] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.856410] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.856422] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.856586] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.856599] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.856826] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.856839] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.856956] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.856968] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.857073] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.857084] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.857265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.857277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.857502] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.857514] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.857625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.857636] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.857863] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.857875] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.857990] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.858002] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.858232] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.858244] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.858474] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.858485] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.858721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.858733] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.858913] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.858925] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.859107] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.859120] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.859285] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.859295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.859552] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.859562] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.859767] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.859777] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.859861] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.859871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.859982] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.859992] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.860123] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.347 [2024-07-15 18:12:49.860133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.347 qpair failed and we were unable to recover it. 00:26:56.347 [2024-07-15 18:12:49.860317] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.860327] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.860433] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.860443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.860610] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.860620] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.860758] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.860768] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.860901] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.860911] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.861020] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.861029] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.861258] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.861268] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.861377] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.861387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.861566] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.861577] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.861774] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.861783] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.861893] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.861903] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.862004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.862013] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.862138] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.862149] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.862258] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.862268] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.862366] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.862375] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.862525] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.862534] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.862639] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.862650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.862804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.862813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.862916] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.862926] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.863195] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.863205] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.863411] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.863421] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.863534] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.863546] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.863717] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.863727] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.863904] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.863914] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.864011] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.864022] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.864193] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.864205] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.864380] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.864392] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.864619] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.864629] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.864805] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.864815] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.864991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.865001] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.865114] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.865124] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.865338] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.865348] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.865467] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.865477] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.865640] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.865650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.865760] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.865770] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.865949] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.865959] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.866154] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.866164] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.866277] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.866286] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.866410] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.866419] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.866516] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.348 [2024-07-15 18:12:49.866525] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.348 qpair failed and we were unable to recover it. 00:26:56.348 [2024-07-15 18:12:49.866748] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.866758] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.867030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.867041] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.867263] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.867273] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.867374] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.867384] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.867495] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.867505] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.867592] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.867602] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.867782] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.867792] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.867908] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.867918] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.868080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.868090] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.868199] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.868208] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.868384] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.868395] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.868566] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.868576] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.868735] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.868745] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.868840] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.868850] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.868938] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.868948] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.869110] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.869120] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.869245] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.869256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.869367] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.869377] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.869454] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.869464] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.869557] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.869567] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.869733] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.869743] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.869901] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.869913] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.870008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.870018] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.870115] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.870125] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.870241] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.870252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.870420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.870430] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.870601] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.870611] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.870789] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.870798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.870939] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.870948] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.871198] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.871208] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.871463] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.871474] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.871644] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.871654] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.871835] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.871845] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.872015] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.872025] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.872133] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.872143] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.872256] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.872267] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.872426] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.872437] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.872670] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.872680] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.872857] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.872867] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.873116] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.349 [2024-07-15 18:12:49.873126] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.349 qpair failed and we were unable to recover it. 00:26:56.349 [2024-07-15 18:12:49.873333] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.873343] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.873498] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.873508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.873738] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.873748] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.873863] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.873873] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.874051] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.874060] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.874160] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.874169] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.874343] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.874353] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.874495] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.874504] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.874755] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.874765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.874959] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.874969] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.875083] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.875092] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.875248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.875258] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.875457] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.875467] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.875633] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.875643] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.875803] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.875813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.876065] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.876075] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.876336] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.876346] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.876539] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.876549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.876798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.876808] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.876995] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.877005] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.877115] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.877124] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.877248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.877260] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.877419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.877429] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.877655] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.877665] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.877865] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.877874] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.878098] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.878108] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.878309] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.878320] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.878492] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.878502] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.878665] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.878675] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.878849] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.878859] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.879081] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.879091] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.879251] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.879262] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.879395] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.879406] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.879632] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.879642] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.879758] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.350 [2024-07-15 18:12:49.879768] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.350 qpair failed and we were unable to recover it. 00:26:56.350 [2024-07-15 18:12:49.879950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.879961] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.880089] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.880099] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.880322] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.880334] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.880568] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.880578] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.880749] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.880759] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.880882] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.880893] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.881050] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.881060] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.881306] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.881316] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.881435] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.881444] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.881641] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.881650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.881850] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.881860] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.882021] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.882030] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.882229] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.882239] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.882486] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.882520] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.882764] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.882780] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.883006] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.883020] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.883184] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.883198] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.883472] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.883487] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.883673] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.883686] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.883921] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.883936] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.884216] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.884236] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.884471] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.884485] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.884605] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.884619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.884743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.884756] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.884991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.885005] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.885274] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.885289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.885532] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.885549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.885829] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.885844] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.886077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.886091] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.886368] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.886383] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.886505] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.886521] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.886704] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.886718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.886845] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.886861] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.886982] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.886996] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.887165] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.887179] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.887414] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.887428] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.887624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.887638] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.887826] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.887840] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.888076] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.351 [2024-07-15 18:12:49.888090] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.351 qpair failed and we were unable to recover it. 00:26:56.351 [2024-07-15 18:12:49.888221] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.888238] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.888434] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.888448] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.888629] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.888643] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.888892] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.888905] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.889192] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.889206] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.889382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.889396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.889524] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.889538] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.889711] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.889725] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.889925] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.889939] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.890179] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.890193] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.890369] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.890384] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.890605] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.890619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.890902] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.890916] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.891090] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.891104] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.891306] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.891318] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.891480] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.891490] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.891693] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.891703] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.891836] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.891846] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.892054] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.892064] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.892289] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.892299] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.892418] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.892428] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.892537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.892547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.892752] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.892762] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.892884] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.892895] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.893138] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.893148] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.893337] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.893348] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.893601] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.893611] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.893798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.893811] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.894086] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.894095] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.894353] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.894363] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.894533] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.894543] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.894656] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.894666] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.894885] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.894895] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.895097] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.895108] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.895283] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.895294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.895401] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.895411] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.895587] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.895597] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.895798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.895808] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.896051] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.896061] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.896316] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.896326] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.352 qpair failed and we were unable to recover it. 00:26:56.352 [2024-07-15 18:12:49.896578] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.352 [2024-07-15 18:12:49.896589] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.896732] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.896744] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.896942] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.896952] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.897210] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.897222] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.897403] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.897414] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.897609] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.897619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.897837] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.897847] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.898045] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.898055] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.898241] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.898252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.898445] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.898455] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.898634] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.898644] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.898979] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.898989] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.899100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.899109] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.899268] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.899278] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.899486] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.899504] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.899713] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.899727] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.899952] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.899966] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.900220] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.900239] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.900479] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.900493] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.900668] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.900682] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.900862] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.900876] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.901004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.901019] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.901274] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.901289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.901522] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.901536] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.901715] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.901728] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.901936] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.901949] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.902192] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.902206] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.902335] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.902352] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.902485] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.902499] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.902682] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.902695] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.902863] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.902877] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.903130] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.903144] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.903396] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.903411] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.903598] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.903613] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.903790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.903805] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.904059] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.904072] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.904254] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.904268] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.904431] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.904445] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.904629] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.904643] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.904897] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.904910] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.905094] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.353 [2024-07-15 18:12:49.905108] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.353 qpair failed and we were unable to recover it. 00:26:56.353 [2024-07-15 18:12:49.905354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.905368] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.905605] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.905619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.905740] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.905754] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.906018] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.906031] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.906215] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.906241] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.906410] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.906424] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.906681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.906695] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.906828] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.906843] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.907076] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.907090] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.907361] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.907377] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.907575] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.907591] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.907786] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.907800] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.907913] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.907927] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.908119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.908132] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.908275] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.908285] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.908505] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.908516] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.908688] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.908699] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.908953] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.908963] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.909216] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.909229] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.909476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.909487] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.909668] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.909678] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.909864] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.909874] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.910039] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.910050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.910231] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.910241] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.910468] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.910478] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.910597] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.910608] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.910880] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.910893] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.911126] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.911136] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.911391] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.911402] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.911574] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.911585] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.911760] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.911770] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.911960] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.911971] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.912182] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.912191] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.912397] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.912407] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.912570] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.912580] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.354 [2024-07-15 18:12:49.912805] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.354 [2024-07-15 18:12:49.912815] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.354 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.913051] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.913061] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.913268] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.913279] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.913505] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.913515] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.913690] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.913700] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.913973] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.913984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.914144] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.914154] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.914388] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.914399] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.914514] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.914524] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.914650] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.914660] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.914894] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.914904] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.915122] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.915132] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.915250] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.915261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.915537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.915548] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.915776] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.915786] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.915951] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.915961] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.916162] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.916172] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.916431] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.916442] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.916678] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.916694] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.916949] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.916964] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.917140] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.917154] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.917332] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.917348] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.917542] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.917556] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.917797] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.917811] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.918033] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.918047] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.918279] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.918293] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.918560] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.918574] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.918690] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.918704] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.918836] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.918850] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.919088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.919102] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.919291] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.919305] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.919556] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.919571] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.919784] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.919798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.920075] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.920089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.920274] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.920289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.920424] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.920438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.920670] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.920684] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.920851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.920864] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.921069] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.921083] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.921336] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.921350] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.921514] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.355 [2024-07-15 18:12:49.921528] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.355 qpair failed and we were unable to recover it. 00:26:56.355 [2024-07-15 18:12:49.921761] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.921775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.922031] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.922044] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.922218] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.922236] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.922411] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.922425] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.922662] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.922676] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.922842] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.922856] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.923094] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.923108] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.923314] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.923329] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.923521] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.923535] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.923720] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.923735] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.923939] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.923953] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.924213] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.924230] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.924395] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.924409] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.924612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.924626] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.924751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.924765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.925030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.925044] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.925277] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.925291] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.925491] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.925508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.925741] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.925755] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.926051] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.926065] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.926302] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.926317] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.926488] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.926502] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.926678] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.926692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.926934] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.926947] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.927131] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.927145] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.927326] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.927341] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.927524] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.927538] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.927720] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.927734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.927902] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.927916] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.928117] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.928131] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.928375] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.928389] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.928616] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.928630] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.928879] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.928893] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.929151] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.929165] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.929396] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.929410] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.929591] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.929605] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.929838] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.929852] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.929979] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.929993] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.930275] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.930289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.930462] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.356 [2024-07-15 18:12:49.930476] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.356 qpair failed and we were unable to recover it. 00:26:56.356 [2024-07-15 18:12:49.930658] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.930672] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.930914] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.930928] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.931118] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.931132] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.931375] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.931389] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.931519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.931532] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.931777] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.931791] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.932055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.932069] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.932328] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.932343] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.932596] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.932610] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.932795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.932809] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.933113] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.933126] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.933354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.933368] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.933545] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.933559] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.933724] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.933737] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.933933] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.933947] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.934125] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.934139] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.934327] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.934342] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.934592] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.934608] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.934788] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.934802] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.935059] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.935073] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.935305] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.935319] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.935505] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.935519] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.935751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.935765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.936091] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.936105] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.936367] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.936382] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.936648] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.936662] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.936931] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.936944] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.937204] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.937218] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.937411] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.937425] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.937620] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.937634] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.937821] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.937836] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.938077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.938091] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.938261] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.938275] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.938406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.938419] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.938658] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.938672] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.938838] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.938852] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.939048] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.939061] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.939247] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.939261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.939430] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.939445] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.939644] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.939658] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.357 [2024-07-15 18:12:49.939782] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.357 [2024-07-15 18:12:49.939795] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.357 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.940046] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.940059] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.940284] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.940299] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.940468] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.940482] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.940612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.940626] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.940812] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.940826] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.940951] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.940965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.941148] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.941163] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.941339] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.941353] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.941485] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.941498] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.941734] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.941748] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.942028] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.942042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.942274] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.942288] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.942497] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.942510] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.942645] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.942659] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.942917] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.942931] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.943108] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.943121] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.943387] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.943404] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.943642] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.943656] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.943824] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.943837] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.944110] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.944124] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.944412] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.944426] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.944550] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.944564] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.944797] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.944811] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.945010] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.945024] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.945219] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.945238] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.945430] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.945444] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.945581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.945595] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.945775] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.945789] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.946069] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.946082] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.946223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.946240] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.946459] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.946473] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.946613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.946627] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.946754] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.946768] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.946987] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.947001] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.947201] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.358 [2024-07-15 18:12:49.947215] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.358 qpair failed and we were unable to recover it. 00:26:56.358 [2024-07-15 18:12:49.947323] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.947337] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.947529] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.947543] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.947780] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.947794] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.948067] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.948081] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.948195] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.948209] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.948387] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.948401] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.948533] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.948547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.948733] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.948747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.948974] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.948989] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.949168] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.949181] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.949358] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.949372] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.949515] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.949528] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.949658] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.949672] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.949853] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.949867] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.949983] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.949997] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.950284] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.950299] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.950416] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.950430] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.950616] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.950630] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.950808] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.950822] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.951079] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.951093] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.951275] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.951289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.951496] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.951513] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.951643] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.951656] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.951932] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.951946] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.952112] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.952126] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.952247] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.952262] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.952482] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.952496] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.952729] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.952743] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.952864] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.952878] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.953114] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.953128] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.953300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.953314] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.953452] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.953466] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.953653] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.953667] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.953803] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.953817] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.954050] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.954064] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.954329] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.954344] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.954522] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.954536] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.954783] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.954797] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.954970] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.954984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.955180] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.955193] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.359 qpair failed and we were unable to recover it. 00:26:56.359 [2024-07-15 18:12:49.955380] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.359 [2024-07-15 18:12:49.955395] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.955600] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.955614] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.955752] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.955766] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.955973] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.955987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.956231] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.956245] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.956518] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.956531] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.956785] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.956798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.956983] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.956997] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.957263] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.957278] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.957500] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.957514] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.957634] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.957648] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.957836] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.957850] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.958086] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.958100] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.958243] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.958257] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.958451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.958464] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.958720] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.958734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.959023] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.959037] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.959299] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.959312] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.959576] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.959591] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.959730] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.959744] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.959972] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.959987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.960223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.960253] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.960436] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.960450] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.960587] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.960600] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.960797] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.960811] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.961072] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.961085] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.961189] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.961202] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.961394] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.961408] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.961641] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.961654] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.961793] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.961807] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.962074] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.962089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.962397] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.962411] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.962598] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.962612] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.962752] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.962767] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.962950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.962965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.963156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.963171] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.963375] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.963390] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.963568] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.963582] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.963684] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.963698] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.963922] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.963936] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.360 [2024-07-15 18:12:49.964132] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.360 [2024-07-15 18:12:49.964146] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.360 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.964324] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.964339] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.964477] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.964490] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.964615] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.964629] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.964763] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.964781] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.965016] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.965031] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.965252] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.965267] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.965385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.965399] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.965589] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.965602] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.965722] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.965735] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.965943] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.965957] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.966127] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.966141] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.966351] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.966366] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.966624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.966639] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.966836] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.966852] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.967106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.967120] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.967360] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.967376] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.967506] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.967520] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.967654] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.967668] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.967854] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.967868] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.968053] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.968067] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.968247] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.968265] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.968479] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.968494] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.968712] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.968727] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.968861] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.968875] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.969163] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.969177] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.969361] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.969375] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.969617] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.969631] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.969763] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.969776] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.970003] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.970017] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.970197] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.970211] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.970429] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.970464] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.970610] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.970626] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.970768] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.970783] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.970976] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.970990] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.971136] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.971150] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.971330] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.971344] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.971486] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.971501] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.971619] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.971633] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.971754] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.971768] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.361 qpair failed and we were unable to recover it. 00:26:56.361 [2024-07-15 18:12:49.971997] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.361 [2024-07-15 18:12:49.972012] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.972134] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.972149] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.972333] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.972349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.972532] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.972548] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.972677] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.972692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.972806] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.972823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.973117] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.973133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.973314] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.973329] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.973456] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.973469] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.973635] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.973649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.973861] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.973875] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.974009] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.974022] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.974256] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.974271] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.974478] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.974492] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.974683] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.974697] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.974811] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.974824] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.975121] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.975135] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.975383] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.975397] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.975588] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.975602] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.975836] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.975850] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.976150] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.976164] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.976352] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.976373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.976498] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.976513] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.976657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.976671] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.976851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.976865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.977118] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.977133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.977353] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.977367] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.977499] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.977513] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.977715] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.977729] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.977931] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.977944] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.978113] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.978128] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.978341] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.978355] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.978475] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.978489] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.978612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.978625] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.978815] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.978830] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.979076] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.979090] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.979258] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.979274] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.979506] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.979521] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.979698] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.979712] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.979920] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.979935] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.980073] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.362 [2024-07-15 18:12:49.980089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.362 qpair failed and we were unable to recover it. 00:26:56.362 [2024-07-15 18:12:49.980281] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.980295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.980467] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.980480] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.980612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.980626] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.980762] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.980775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.980968] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.980981] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.981187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.981201] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.981391] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.981406] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.981607] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.981637] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.981784] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.981799] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.982034] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.982048] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.982272] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.982287] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.982424] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.982438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.982630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.982644] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.982848] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.982863] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.983036] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.983050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.983161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.983175] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.983346] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.983361] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.983560] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.983573] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.983798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.983812] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.984023] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.984037] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.984156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.984170] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.984385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.984400] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.984542] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.984556] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.984680] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.984694] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.984809] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.984823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.985059] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.985073] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.985247] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.985262] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.985389] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.985403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.985621] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.985635] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.985753] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.985767] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.985941] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.985954] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.986143] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.986156] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.986407] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.986424] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.986606] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.986620] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.986746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.986762] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.363 qpair failed and we were unable to recover it. 00:26:56.363 [2024-07-15 18:12:49.986971] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.363 [2024-07-15 18:12:49.986984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.987152] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.987165] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.987429] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.987444] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.987633] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.987647] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.987860] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.987874] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.988155] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.988169] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.988383] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.988397] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.988637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.988651] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.988830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.988844] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.989167] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.989180] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.989354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.989369] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.989549] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.989563] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.989776] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.989790] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.989911] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.989925] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.990035] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.990049] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.990315] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.990330] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.990518] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.990532] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.990675] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.990688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.990867] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.990881] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.991105] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.991119] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.991352] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.991366] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.991609] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.991624] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.991831] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.991845] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.992104] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.992118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.992290] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.992304] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.992491] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.992506] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.992689] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.992702] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.992928] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.992942] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.993200] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.993214] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.993344] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.993358] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.993526] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.993541] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.993738] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.993753] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.993886] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.993900] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.994149] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.994163] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.994365] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.994379] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.994575] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.994588] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.994726] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.994740] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.994970] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.994984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.995102] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.995116] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.995294] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.995311] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.364 qpair failed and we were unable to recover it. 00:26:56.364 [2024-07-15 18:12:49.995515] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.364 [2024-07-15 18:12:49.995529] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.995683] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.995697] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.995816] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.995830] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.996055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.996069] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.996250] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.996265] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.996455] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.996469] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.996600] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.996614] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.996792] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.996806] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.996993] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.997008] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.997246] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.997261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.997393] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.997407] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.997599] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.997614] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.997755] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.997769] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.998050] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.998064] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.998241] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.998256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.998457] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.998471] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.998612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.998626] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.998798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.998811] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.999016] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.999030] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.999254] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.999269] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.999457] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.999471] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:49.999728] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:49.999742] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:50.000049] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:50.000063] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:50.000327] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:50.000341] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:50.000526] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:50.000540] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:50.000664] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:50.000679] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:50.000895] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:50.000910] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:50.001023] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:50.001038] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:50.001301] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:50.001315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:50.001432] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:50.001446] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:50.001627] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:50.001641] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:50.001780] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:50.001794] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:50.001986] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:50.002000] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:50.002205] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:50.002219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:50.002400] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:50.002414] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:50.002550] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:50.002564] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:50.002798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:50.002812] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:50.003052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:50.003066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:50.003339] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:50.003353] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.365 qpair failed and we were unable to recover it. 00:26:56.365 [2024-07-15 18:12:50.003540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.365 [2024-07-15 18:12:50.003557] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.003761] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.003775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.003903] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.003917] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.004104] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.004118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.004289] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.004304] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.004428] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.004443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.004578] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.004593] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.004723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.004737] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.004877] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.004891] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.005067] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.005081] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.005308] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.005323] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.005451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.005464] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.005587] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.005601] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.005704] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.005718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.005859] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.005873] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.006070] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.006084] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.006201] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.006215] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.006366] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.006380] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.006511] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.006525] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.006712] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.006727] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.006857] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.006871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.007100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.007115] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.007378] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.007393] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.007533] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.007547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.007685] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.007699] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.007835] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.007848] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.008026] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.008039] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.008228] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.008242] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.008416] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.008429] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.008601] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.008614] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.008803] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.008817] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.009019] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.009033] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.009303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.009317] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.009433] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.009446] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.009571] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.009585] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.009702] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.009716] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.009874] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.009888] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.010012] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.010026] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.010210] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.010230] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.010448] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.010462] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.010672] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.010688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.366 [2024-07-15 18:12:50.010875] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.366 [2024-07-15 18:12:50.010889] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.366 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.011119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.011133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.011331] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.011348] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.011576] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.011591] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.011779] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.011793] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.012046] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.012060] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.012245] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.012259] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.012436] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.012450] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.012589] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.012603] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.012837] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.012851] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.013085] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.013099] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.013291] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.013305] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.013426] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.013439] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.013580] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.013594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.013770] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.013784] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.014001] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.014015] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.014259] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.014273] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.014415] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.014430] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.014655] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.014669] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.014974] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.014988] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.015241] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.015256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.015391] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.015405] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.015593] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.015607] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.015707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.015721] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.015839] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.015853] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.016037] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.016051] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.016280] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.016295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.016433] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.016446] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.016649] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.016663] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.016835] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.016849] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.017132] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.017146] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.017283] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.017298] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.017493] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.017508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.017732] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.017746] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.017937] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.017950] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.018147] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.018160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.018274] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.018289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.018425] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.018440] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.367 qpair failed and we were unable to recover it. 00:26:56.367 [2024-07-15 18:12:50.018580] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.367 [2024-07-15 18:12:50.018594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.018773] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.018789] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.018958] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.018971] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.019149] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.019163] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.019333] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.019348] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.019522] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.019536] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.019769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.019782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.020014] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.020029] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.020210] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.020236] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.020447] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.020461] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.020576] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.020589] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.020727] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.020741] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.021065] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.021078] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.021256] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.021271] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.021460] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.021474] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.021587] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.021601] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.021744] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.021759] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.021882] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.021895] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.022133] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.022147] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.022302] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.022316] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.022437] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.022452] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.022542] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.022556] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.022723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.022737] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.022944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.022958] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.023093] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.023108] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.023355] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.023369] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.023502] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.023517] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.023710] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.023724] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.024011] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.024026] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.024208] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.024221] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.024449] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.024463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.024603] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.024617] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.024737] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.024751] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.024941] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.024955] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.025191] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.025205] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.025395] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.025418] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.025575] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.025593] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.027887] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.027923] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.028068] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.028084] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.028287] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.028303] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.028540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.028555] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.028692] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.028711] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.028827] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.028840] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.028969] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.368 [2024-07-15 18:12:50.028982] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.368 qpair failed and we were unable to recover it. 00:26:56.368 [2024-07-15 18:12:50.029111] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.369 [2024-07-15 18:12:50.029126] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.369 qpair failed and we were unable to recover it. 00:26:56.369 [2024-07-15 18:12:50.029263] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.369 [2024-07-15 18:12:50.029278] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.369 qpair failed and we were unable to recover it. 00:26:56.369 [2024-07-15 18:12:50.029500] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.369 [2024-07-15 18:12:50.029515] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.369 qpair failed and we were unable to recover it. 00:26:56.369 [2024-07-15 18:12:50.029636] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.369 [2024-07-15 18:12:50.029649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.369 qpair failed and we were unable to recover it. 00:26:56.369 [2024-07-15 18:12:50.029827] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.369 [2024-07-15 18:12:50.029841] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.369 qpair failed and we were unable to recover it. 00:26:56.652 [2024-07-15 18:12:50.029967] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.652 [2024-07-15 18:12:50.029981] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.652 qpair failed and we were unable to recover it. 00:26:56.652 [2024-07-15 18:12:50.030098] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.652 [2024-07-15 18:12:50.030112] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.652 qpair failed and we were unable to recover it. 00:26:56.652 [2024-07-15 18:12:50.030240] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.652 [2024-07-15 18:12:50.030254] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.652 qpair failed and we were unable to recover it. 00:26:56.652 [2024-07-15 18:12:50.030374] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.652 [2024-07-15 18:12:50.030388] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.652 qpair failed and we were unable to recover it. 00:26:56.652 [2024-07-15 18:12:50.030508] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.652 [2024-07-15 18:12:50.030522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.652 qpair failed and we were unable to recover it. 00:26:56.652 [2024-07-15 18:12:50.030624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.652 [2024-07-15 18:12:50.030638] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.652 qpair failed and we were unable to recover it. 00:26:56.652 [2024-07-15 18:12:50.030798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.652 [2024-07-15 18:12:50.030812] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.652 qpair failed and we were unable to recover it. 00:26:56.652 [2024-07-15 18:12:50.030940] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.652 [2024-07-15 18:12:50.030955] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.652 qpair failed and we were unable to recover it. 00:26:56.652 [2024-07-15 18:12:50.031080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.652 [2024-07-15 18:12:50.031095] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.652 qpair failed and we were unable to recover it. 00:26:56.652 [2024-07-15 18:12:50.031206] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.652 [2024-07-15 18:12:50.031220] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.652 qpair failed and we were unable to recover it. 00:26:56.652 [2024-07-15 18:12:50.031349] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.652 [2024-07-15 18:12:50.031364] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.652 qpair failed and we were unable to recover it. 00:26:56.652 [2024-07-15 18:12:50.031535] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.652 [2024-07-15 18:12:50.031551] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.652 qpair failed and we were unable to recover it. 00:26:56.652 [2024-07-15 18:12:50.031677] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.652 [2024-07-15 18:12:50.031692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.652 qpair failed and we were unable to recover it. 00:26:56.652 [2024-07-15 18:12:50.031964] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.652 [2024-07-15 18:12:50.031979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.652 qpair failed and we were unable to recover it. 00:26:56.652 [2024-07-15 18:12:50.032214] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.652 [2024-07-15 18:12:50.032234] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.652 qpair failed and we were unable to recover it. 00:26:56.652 [2024-07-15 18:12:50.032401] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.652 [2024-07-15 18:12:50.032415] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.652 qpair failed and we were unable to recover it. 00:26:56.652 [2024-07-15 18:12:50.032549] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.652 [2024-07-15 18:12:50.032562] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.652 qpair failed and we were unable to recover it. 00:26:56.652 [2024-07-15 18:12:50.032687] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.652 [2024-07-15 18:12:50.032701] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.652 qpair failed and we were unable to recover it. 00:26:56.652 [2024-07-15 18:12:50.032891] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.652 [2024-07-15 18:12:50.032905] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.033015] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.033029] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.033248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.033263] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.033447] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.033461] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.033559] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.033573] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.033778] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.033792] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.033964] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.033979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.034240] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.034254] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.034472] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.034486] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.034674] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.034688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.034818] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.034832] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.034966] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.034981] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.035121] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.035135] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.035336] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.035351] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.035463] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.035480] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.035648] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.035663] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.035910] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.035923] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.036033] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.036047] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.036156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.036170] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.036367] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.036382] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.036571] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.036585] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.036708] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.036722] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.036844] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.036858] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.037039] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.037053] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.037239] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.037253] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.037377] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.037391] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.037512] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.037525] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.037733] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.037746] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.037939] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.037953] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.038161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.038176] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.038305] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.038319] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.038492] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.038506] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.038685] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.038699] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.038821] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.038835] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.039057] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.039071] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.039274] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.039288] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.039416] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.039430] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.039552] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.039566] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.039678] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.039692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.039824] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.039838] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.040065] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.040079] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.040322] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.040336] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.040519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.040533] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.040722] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.040736] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.040924] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.040938] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.041116] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.041130] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.041251] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.041265] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.041435] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.041449] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.041573] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.041587] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.041760] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.041773] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.041906] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.041919] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.042179] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.042192] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.042314] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.042329] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.042440] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.042453] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.042571] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.042587] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.042717] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.042731] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.042938] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.042952] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.043120] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.043134] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.043352] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.043366] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.043496] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.043510] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.043643] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.043657] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.043787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.043800] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.043973] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.043987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.044215] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.044232] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.044415] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.044429] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.044598] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.044612] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.653 [2024-07-15 18:12:50.044743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.653 [2024-07-15 18:12:50.044757] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.653 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.044887] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.044901] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.045029] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.045043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.045279] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.045294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.045483] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.045497] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.045618] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.045632] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.045772] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.045785] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.046086] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.046100] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.046373] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.046387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.046504] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.046518] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.046700] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.046714] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.046888] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.046903] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.047084] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.047097] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.047300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.047315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.047561] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.047575] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.047766] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.047794] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.048099] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.048122] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.048314] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.048324] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.048550] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.048561] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.048860] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.048871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.049052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.049063] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.049181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.049192] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.049305] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.049316] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.049428] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.049438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.049664] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.049675] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.049893] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.049903] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.050129] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.050139] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.050296] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.050307] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.050577] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.050590] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.050753] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.050762] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.050975] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.050985] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.051103] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.051113] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.051362] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.051372] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.051538] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.051549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.051718] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.051728] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.051960] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.051970] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.052150] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.052160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.052333] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.052343] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.052574] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.052584] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.052759] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.052769] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.052981] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.052991] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.053220] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.053234] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.053425] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.053435] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.053669] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.053679] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.053799] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.053810] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.054084] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.054095] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.054312] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.054322] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.054502] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.054512] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.054689] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.054700] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.054914] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.054925] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.055220] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.055236] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.055385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.055396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.055512] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.055522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.055698] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.055709] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.055832] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.055842] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.055993] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.056010] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.056297] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.056311] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.056500] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.056514] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.056699] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.056713] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.057005] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.057020] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.057275] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.057290] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.057512] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.057526] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.057654] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.057668] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.057804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.654 [2024-07-15 18:12:50.057819] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.654 qpair failed and we were unable to recover it. 00:26:56.654 [2024-07-15 18:12:50.058000] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.058013] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.058191] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.058204] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.058319] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.058335] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.058525] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.058538] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.058718] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.058734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.058842] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.058856] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.059032] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.059047] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.059281] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.059295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.059540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.059554] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.059788] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.059802] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.059930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.059944] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.060196] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.060210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.060321] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.060335] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.060520] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.060533] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.060666] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.060680] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.060923] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.060937] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.061162] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.061175] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.061391] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.061405] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.061613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.061627] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.061759] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.061773] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.061905] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.061919] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.062128] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.062142] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.062317] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.062331] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.062568] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.062581] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.062754] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.062768] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.063032] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.063046] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.063234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.063248] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.063401] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.063415] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.063581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.063595] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.063774] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.063787] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.064012] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.064025] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.064209] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.064220] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.064421] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.064432] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.064608] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.064619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.064751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.064760] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.064946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.064956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.065161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.065171] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.065303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.065314] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.065486] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.065496] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.065721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.065731] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.065841] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.065851] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.065974] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.065984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.066161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.066171] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.066338] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.066349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.066547] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.066560] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.066675] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.066685] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.066788] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.066798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.066919] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.066929] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.067110] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.067120] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.067246] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.067256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.067384] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.067394] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.067493] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.067503] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.067618] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.067628] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.067734] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.067745] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.655 qpair failed and we were unable to recover it. 00:26:56.655 [2024-07-15 18:12:50.067851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.655 [2024-07-15 18:12:50.067861] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.067974] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.067984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.068149] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.068160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.068390] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.068400] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.068502] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.068512] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.068604] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.068614] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.068742] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.068752] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.068871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.068880] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.069055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.069066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.069156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.069166] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.069280] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.069290] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.069462] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.069472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.069565] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.069574] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.069691] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.069701] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.069802] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.069812] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.069973] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.069983] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.070100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.070111] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.070213] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.070237] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.070339] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.070349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.070452] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.070463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.070537] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.070547] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.070731] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.070742] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.070909] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.070919] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.071035] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.071045] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.071242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.071253] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.071362] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.071372] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.071476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.071485] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.071591] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.071601] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.071784] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.071794] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.071963] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.071974] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.072091] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.072103] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.072279] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.072289] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.072396] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.072407] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.072505] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.072516] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.072630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.072641] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.072734] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.072744] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.072855] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.072866] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.072996] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.073006] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.073181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.073192] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.073290] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.073300] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.073553] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.073563] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.073667] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.073678] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.073803] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.073813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.073918] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.073928] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.074052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.074062] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.074162] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.074172] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.074269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.074279] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.074456] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.074466] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.074573] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.074583] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.074724] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.074734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.074834] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.074844] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.075097] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.075108] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.075298] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.075308] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.075441] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.075451] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.075543] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.075553] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.075659] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.075669] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.075768] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.075778] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.075887] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.075898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.075984] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.075994] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.076156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.076167] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.076259] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.656 [2024-07-15 18:12:50.076270] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.656 qpair failed and we were unable to recover it. 00:26:56.656 [2024-07-15 18:12:50.076374] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.076384] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.076467] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.076477] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.076642] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.076652] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.076829] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.076838] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.076944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.076954] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.077138] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.077148] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.077243] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.077253] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.077410] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.077421] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.077609] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.077619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.077703] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.077714] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.077821] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.077831] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.077924] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.077934] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.078015] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.078025] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.078144] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.078154] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.078265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.078275] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.078363] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.078374] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.078465] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.078475] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.078573] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.078583] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.078651] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.078661] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.078895] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.078905] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.079000] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.079010] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.079108] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.079118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.079281] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.079291] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.079404] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.079414] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.079535] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.079545] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.079804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.079814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.079919] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.079929] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.080020] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.080030] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.080137] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.080147] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.080322] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.080332] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.080507] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.080516] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.080627] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.080637] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.080730] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.080740] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.080834] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.080844] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.080949] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.080960] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.081072] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.081083] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.081208] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.081229] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.081322] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.081336] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.081495] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.081509] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.081612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.081626] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.081771] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.081785] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.081953] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.081967] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.082076] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.082089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.082172] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.082186] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.082295] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.082310] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.082420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.082433] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.082541] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.082555] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.082737] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.082751] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.082860] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.082874] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.082961] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.082974] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.083085] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.083099] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.083254] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.083269] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.083385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.083399] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.083508] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.083522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.083633] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.083646] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.083761] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.083775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.083888] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.083902] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.084110] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.084124] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.657 [2024-07-15 18:12:50.084285] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.657 [2024-07-15 18:12:50.084298] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.657 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.084426] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.084440] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.084559] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.084573] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.084683] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.084697] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.084869] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.084884] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.084989] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.085005] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.085111] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.085125] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.085335] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.085349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.085469] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.085483] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.085668] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.085682] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.085780] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.085793] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.085904] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.085919] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.086087] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.086101] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.086209] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.086223] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.086345] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.086358] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.086464] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.086478] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.086589] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.086604] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.086689] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.086703] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.086885] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.086902] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.087071] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.087086] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.087190] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.087203] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.087374] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.087388] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.087492] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.087506] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.087608] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.087622] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.087798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.087812] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.087919] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.087933] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.088108] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.088122] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.088237] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.088251] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.088373] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.088387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.088503] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.088517] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.088628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.088642] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.088744] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.088758] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.088884] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.088898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.089008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.089022] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.089114] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.089128] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.089236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.089250] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.089371] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.089385] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.089504] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.089519] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.089623] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.089636] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.089751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.089765] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.089858] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.089872] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.089987] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.090002] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.090102] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.090115] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.090229] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.090243] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.090324] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.090338] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.090454] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.090468] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.090570] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.090584] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.090695] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.090708] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.090787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.090801] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.090912] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.090926] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.091041] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.091055] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.091169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.091182] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.091354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.091369] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.091459] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.091473] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.658 [2024-07-15 18:12:50.091570] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.658 [2024-07-15 18:12:50.091584] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.658 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.091690] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.091703] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.091805] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.091818] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.091920] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.091934] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.092032] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.092049] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.092242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.092257] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.092373] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.092387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.092487] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.092500] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.092604] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.092618] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.092778] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.092792] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.093048] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.093061] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.093177] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.093191] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.093301] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.093316] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.093495] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.093510] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.093625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.093638] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.093811] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.093826] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.093935] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.093949] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.094063] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.094077] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.094256] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.094271] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.094374] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.094388] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.094487] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.094500] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.094611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.094624] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.094804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.094818] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.094999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.095013] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.095141] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.095154] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.095264] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.095279] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.095388] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.095403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.095516] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.095530] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.095631] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.095645] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.095752] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.095766] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.095867] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.095880] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.095981] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.095996] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.096106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.096120] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.096342] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.096356] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.096471] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.096485] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.096656] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.096669] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.096851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.096865] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.097034] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.097048] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.097157] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.097171] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.097273] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.097287] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.097406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.097420] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.097561] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.097575] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.097684] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.097698] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.097823] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.097837] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.097994] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.098010] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.098127] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.098141] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.098260] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.098274] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.098444] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.098458] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.098554] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.098568] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.098733] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.098747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.098872] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.098886] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.099056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.099069] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.099179] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.099193] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.099320] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.099334] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.099500] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.099514] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.099692] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.099707] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.099817] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.099831] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.099927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.099940] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.100052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.100066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.100250] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.100264] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.100373] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.100386] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.100575] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.100589] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.100683] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.100697] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.659 [2024-07-15 18:12:50.100882] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.659 [2024-07-15 18:12:50.100897] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.659 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.101107] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.101122] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.101301] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.101315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.101419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.101434] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.101604] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.101618] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.101742] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.101756] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.101871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.101885] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.101980] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.101993] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.102110] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.102123] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.102251] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.102266] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.102369] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.102383] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.102465] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.102478] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.102593] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.102608] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.102706] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.102719] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.102893] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.102907] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.103006] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.103020] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.103159] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.103173] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.103294] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.103309] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.103412] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.103425] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.103534] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.103548] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.103735] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.103749] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.103913] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.103929] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.104116] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.104130] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.104255] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.104270] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.104357] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.104371] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.104485] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.104499] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.104607] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.104621] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.104741] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.104754] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.104865] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.104879] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.104995] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.105009] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.105128] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.105142] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.105254] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.105268] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.105348] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.105362] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.105482] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.105495] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.105595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.105609] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.105713] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.105727] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.105825] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.105839] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.105954] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.105968] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.106090] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.106104] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.106258] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.106272] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.106442] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.106456] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.106605] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.106618] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.106731] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.106745] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.106929] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.106943] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.107115] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.107129] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.107364] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.107378] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.107495] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.107511] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.107617] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.107631] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.107744] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.107758] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.107867] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.107881] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.107958] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.107972] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.108146] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.108160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.108242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.108257] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.108355] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.108369] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.108479] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.108493] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.108612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.108628] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.108727] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.108741] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.108913] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.108926] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.109038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.109048] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.109142] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.109152] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.109386] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.109396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.109465] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.109478] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.109613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.109624] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.109804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.109814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.109918] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.660 [2024-07-15 18:12:50.109928] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.660 qpair failed and we were unable to recover it. 00:26:56.660 [2024-07-15 18:12:50.110031] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.110041] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.110151] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.110161] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.110323] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.110334] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.110420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.110430] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.110589] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.110600] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.110765] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.110775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.110946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.110956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.111053] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.111063] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.111165] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.111175] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.111339] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.111349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.111450] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.111460] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.111634] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.111644] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.111803] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.111814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.111998] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.112008] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.112108] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.112118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.112234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.112245] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.112355] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.112366] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.112466] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.112476] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.112594] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.112605] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.112697] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.112707] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.112811] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.112821] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.112982] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.112992] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.113183] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.113193] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.113303] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.113314] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.113409] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.113419] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.113513] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.113523] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.113683] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.113693] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.113800] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.113810] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.113903] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.113914] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.114032] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.114042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.114152] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.114162] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.114280] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.114291] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.114407] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.114417] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.114524] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.114534] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.114634] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.114644] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.114821] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.114831] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.114952] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.114965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.115065] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.115075] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.115182] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.115192] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.115347] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.115357] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.115467] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.115477] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.115639] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.115649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.115759] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.115769] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.115870] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.115880] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.115973] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.115983] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.116098] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.116109] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.116270] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.116281] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.116386] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.116396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.116625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.116636] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.116797] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.116807] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.116907] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.116917] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.117032] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.117042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.117131] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.117141] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.117251] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.117262] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.117383] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.117393] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.117583] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.117593] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.117845] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.117855] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.661 qpair failed and we were unable to recover it. 00:26:56.661 [2024-07-15 18:12:50.117955] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.661 [2024-07-15 18:12:50.117965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.118071] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.118081] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.118196] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.118206] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.118435] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.118445] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.118627] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.118637] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.118797] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.118807] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.118917] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.118927] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.119015] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.119025] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.119215] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.119228] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.119490] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.119501] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.119662] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.119672] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.119772] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.119782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.119976] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.119987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.120147] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.120157] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.120276] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.120286] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.120458] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.120468] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.120651] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.120661] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.120835] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.120846] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.121072] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.121082] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.121256] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.121269] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.121378] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.121388] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.121485] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.121495] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.121586] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.121596] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.121709] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.121720] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.121815] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.121826] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.121996] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.122006] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.122209] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.122220] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.122323] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.122333] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.122506] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.122517] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.122728] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.122738] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.122856] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.122866] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.123033] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.123042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.123148] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.123159] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.123278] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.123288] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.123402] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.123413] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.123519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.123529] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.123689] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.123699] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.123860] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.123870] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.124097] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.124107] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.124198] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.124209] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.124338] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.124348] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.124476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.124487] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.124584] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.124594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.124762] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.124772] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.124886] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.124896] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.125002] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.125012] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.125179] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.125189] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.125312] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.125322] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.125486] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.125496] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.125599] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.125609] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.125713] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.125723] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.125890] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.125901] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.126034] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.126044] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.126155] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.126165] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.126304] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.126315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.662 [2024-07-15 18:12:50.126420] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.662 [2024-07-15 18:12:50.126429] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.662 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.126522] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.126533] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.126625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.126635] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.126748] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.126759] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.126922] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.126935] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.127045] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.127055] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.127151] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.127161] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.127331] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.127342] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.127434] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.127444] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.127551] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.127561] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.127660] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.127670] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.127776] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.127786] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.127944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.127955] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.128139] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.128149] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.128266] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.128277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.128390] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.128401] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.128576] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.128585] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.128695] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.128705] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.128807] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.128816] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.128961] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.128971] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.129071] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.129081] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.129241] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.129251] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.129344] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.129354] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.129516] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.129526] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.129630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.129640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.129801] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.129812] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.129930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.129940] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.130064] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.130074] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.130171] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.130181] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.130282] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.130293] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.130464] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.130474] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.130569] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.130580] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.130673] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.130683] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.130779] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.130789] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.130881] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.130891] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.130990] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.131000] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.131093] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.131102] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.131194] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.131204] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.131302] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.131314] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.131473] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.131484] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.131595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.131606] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.131792] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.131802] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.131943] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.131954] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.132123] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.132134] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.132369] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.132382] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.132499] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.132509] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.132671] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.132681] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.132778] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.132788] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.132883] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.132893] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.132998] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.133009] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.133077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.133087] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.133192] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.133203] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.133280] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.133291] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.133386] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.133396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.133517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.133527] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.133707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.133717] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.663 [2024-07-15 18:12:50.133948] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.663 [2024-07-15 18:12:50.133958] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.663 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.134114] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.134124] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.134298] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.134309] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.134403] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.134413] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.134503] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.134513] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.134674] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.134684] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.134787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.134796] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.134970] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.134980] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.135077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.135088] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.135182] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.135192] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.135300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.135310] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.135409] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.135419] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.135624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.135634] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.135733] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.135742] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.135849] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.135859] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.136051] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.136083] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.136206] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.136222] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.136470] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.136484] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.136652] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.136666] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.136795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.136809] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.136930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.136945] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.137096] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.137110] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.137223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.137241] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.137415] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.137429] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.137529] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.137543] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.137679] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.137693] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.137795] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.137809] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.138021] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.138035] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.138210] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.138230] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.138445] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.138460] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.138632] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.138646] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.138755] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.138769] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.138937] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.138950] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.139065] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.139079] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.139183] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.139196] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.139310] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.139325] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.139438] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.139452] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.139565] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.139579] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.139697] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.139711] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.139801] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.139814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.139982] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.139996] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.140118] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.140132] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.140322] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.140340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.140509] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.140524] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.140644] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.140658] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.140765] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.140779] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.140897] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.140910] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.141096] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.141110] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.141233] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.141247] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.141359] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.141372] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.141474] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.141487] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.141614] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.141629] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.141743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.141757] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.141862] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.141876] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.141980] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.141994] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.664 qpair failed and we were unable to recover it. 00:26:56.664 [2024-07-15 18:12:50.142120] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.664 [2024-07-15 18:12:50.142134] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.142325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.142339] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.142458] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.142472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.142595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.142609] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.142733] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.142747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.142924] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.142939] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.143119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.143133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.143304] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.143318] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.143430] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.143443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.143610] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.143624] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.143793] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.143807] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.143972] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.143985] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.144092] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.144104] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.144206] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.144216] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.144313] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.144327] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.144428] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.144439] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.144609] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.144619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.144715] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.144725] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.144968] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.144978] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.145077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.145087] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.145193] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.145204] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.145305] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.145316] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.145482] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.145492] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.145603] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.145613] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.145723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.145733] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.145858] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.145868] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.145968] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.145978] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.146067] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.146077] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.146183] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.146193] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.146324] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.146334] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.146430] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.146440] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.146612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.146623] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.146856] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.146866] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.146968] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.146978] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.147083] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.147093] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.147210] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.147220] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.147406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.147416] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.147596] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.147606] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.147706] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.147716] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.147812] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.147821] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.147929] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.147939] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.148041] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.148052] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.148167] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.148176] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.148285] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.148296] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.148402] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.148412] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.148506] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.148516] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.148665] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.148674] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.148763] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.148773] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.148878] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.148889] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.148981] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.148991] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.149146] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.149155] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.149316] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.149327] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.149419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.149429] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.149594] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.149603] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.149697] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.149708] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.149809] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.149819] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.149907] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.149917] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.149983] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.149993] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.150184] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.150194] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.150357] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.150367] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.150455] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.150465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.150624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.150634] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.665 [2024-07-15 18:12:50.150729] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.665 [2024-07-15 18:12:50.150738] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.665 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.150833] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.150843] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.150963] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.150972] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.151156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.151166] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.151274] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.151284] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.151460] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.151470] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.151565] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.151575] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.151721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.151737] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.151831] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.151841] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.151937] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.151947] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.152062] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.152072] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.152178] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.152189] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.152305] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.152316] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.152397] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.152407] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.152513] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.152524] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.152634] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.152644] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.152739] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.152749] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.152845] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.152855] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.153037] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.153047] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.153151] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.153161] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.153267] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.153277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.153368] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.153378] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.153475] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.153485] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.153581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.153591] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.153682] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.153692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.153810] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.153821] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.153920] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.153930] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.154090] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.154100] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.154201] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.154211] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.154401] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.154411] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.154514] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.154524] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.154644] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.154654] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.154766] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.154778] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.154881] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.154891] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.155001] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.155013] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.155106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.155117] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.155295] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.155306] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.155399] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.155410] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.155499] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.155509] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.155617] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.155627] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.155735] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.155746] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.155850] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.155859] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.155972] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.155983] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.156153] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.156163] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.156274] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.156284] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.156378] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.156388] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.156499] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.156509] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.156611] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.156621] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.156804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.156814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.156906] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.156916] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.157044] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.157054] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.157164] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.157175] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.157343] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.157354] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.157525] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.157536] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.157632] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.157642] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.157817] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.157828] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.666 [2024-07-15 18:12:50.158023] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.666 [2024-07-15 18:12:50.158032] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.666 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.158138] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.158148] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.158339] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.158349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.158447] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.158457] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.158634] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.158645] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.158757] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.158767] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.158866] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.158877] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.158966] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.158976] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.159065] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.159076] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.159169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.159179] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.159279] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.159290] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.159382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.159392] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.159572] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.159582] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.159680] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.159689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.159794] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.159804] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.159969] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.159979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.160091] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.160102] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.160261] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.160271] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.160386] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.160396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.160522] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.160532] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.160693] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.160704] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.160816] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.160827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.160919] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.160929] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.161033] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.161043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.161211] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.161221] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.161323] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.161333] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.161507] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.161517] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.161607] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.161618] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.161790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.161801] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.161906] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.161916] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.162028] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.162038] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.162132] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.162142] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.162261] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.162274] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.162369] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.162379] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.162484] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.162494] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.162587] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.162598] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.162690] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.162699] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.162792] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.162802] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.162927] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.162937] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.163032] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.163042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.163161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.163171] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.163264] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.163274] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.163440] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.163450] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.163648] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.163659] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.163745] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.163755] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.163958] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.163968] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.164242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.164252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.164367] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.164377] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.164492] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.164502] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.164597] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.164608] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.164705] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.164716] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.164878] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.164888] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.165056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.165067] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.165170] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.165181] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.165296] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.165307] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.165472] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.165483] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.165581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.165594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.165763] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.165774] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.165871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.165881] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.165980] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.165991] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.166169] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.166179] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.166285] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.166295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.166451] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.166462] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.166622] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.166632] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.166724] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.166734] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.166839] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.667 [2024-07-15 18:12:50.166849] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.667 qpair failed and we were unable to recover it. 00:26:56.667 [2024-07-15 18:12:50.167091] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.167101] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.167197] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.167207] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.167321] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.167331] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.167500] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.167510] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.167689] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.167701] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.167792] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.167803] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.167913] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.167923] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.168098] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.168109] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.168284] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.168295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.168416] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.168428] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.168527] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.168538] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.168652] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.168663] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.168745] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.168756] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.168847] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.168858] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.169027] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.169037] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.169128] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.169138] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.169372] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.169383] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.169461] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.169478] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.169584] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.169600] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.169721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.169735] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.169969] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.169984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.170086] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.170101] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.170237] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.170252] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.170366] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.170382] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.170499] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.170513] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.170616] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.170630] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.170742] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.170756] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.170867] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.170882] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.170991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.171006] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.171129] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.171141] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.171318] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.171331] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.171444] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.171455] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.171556] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.171566] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.171660] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.171670] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.171778] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.171788] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.171911] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.171921] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.172022] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.172033] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.172298] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.172309] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.172414] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.172424] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.172526] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.172536] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.172608] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.172619] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.172716] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.172726] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.172832] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.172842] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.173047] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.173058] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.173220] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.173242] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.173354] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.173365] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.173599] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.173610] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.173769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.173779] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.173942] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.173952] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.174176] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.174187] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.174348] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.174360] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.174468] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.174478] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.174648] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.174657] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.174753] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.174764] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.174935] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.174945] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.175113] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.175124] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.175243] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.175253] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.175462] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.175478] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.175604] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.175618] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.175735] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.175749] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.668 [2024-07-15 18:12:50.175858] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.668 [2024-07-15 18:12:50.175872] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.668 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.175992] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.176007] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.176109] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.176124] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.176235] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.176249] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.176362] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.176377] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.176508] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.176522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.176595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.176609] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.176787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.176802] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.176893] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.176907] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.177009] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.177023] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.177124] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.177141] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.177292] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.177307] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.177519] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.177534] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.177706] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.177720] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.177838] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.177852] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.178024] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.178038] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.178251] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.178266] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.178444] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.178458] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.178601] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.178615] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.178723] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.178737] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.178926] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.178940] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.179012] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.179026] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.179182] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.179196] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.179313] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.179327] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.179459] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.179474] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.179708] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.179722] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.179827] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.179842] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.179958] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.179973] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.180164] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.180178] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.180284] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.180298] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.180401] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.180415] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.180544] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.180559] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.180678] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.180692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.180808] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.180822] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.180925] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.180939] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.181025] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.181039] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.181141] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.181156] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.181293] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.181306] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.181422] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.181433] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.181536] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.181546] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.181699] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.181710] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.181858] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.181869] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.182033] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.182042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.182160] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.182171] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.182244] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.182255] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.182425] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.182439] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.182517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.182527] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.182755] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.182766] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.182861] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.182872] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.182984] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.182994] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.183123] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.183136] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.183368] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.183380] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.183492] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.183503] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.183674] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.183685] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.669 [2024-07-15 18:12:50.183796] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.669 [2024-07-15 18:12:50.183807] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.669 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.183904] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.183914] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.184064] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.184075] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.184204] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.184215] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.184444] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.184456] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.184559] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.184571] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.184681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.184692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.184786] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.184797] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.184970] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.184981] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.185110] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.185121] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.185203] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.185213] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.185397] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.185408] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.185581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.185592] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.185700] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.185711] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.185810] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.185821] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.185990] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.186001] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.186109] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.186118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.186214] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.186229] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.186335] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.186346] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.186446] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.186456] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.186631] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.186641] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.186817] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.186827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.186919] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.186930] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.187055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.187071] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.187190] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.187204] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.187321] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.187336] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.187470] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.187485] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.187583] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.187598] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.187702] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.187716] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.187909] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.187924] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.188054] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.188068] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.188305] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.188320] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.188489] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.188504] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.188593] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.188608] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.188775] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.188790] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.188908] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.188922] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.189028] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.189042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.189164] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.189178] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.189286] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.189302] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.189488] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.189501] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.189607] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.189621] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.189732] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.189746] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.189858] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.189873] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.190039] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.190054] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.190298] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.190313] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.190480] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.190494] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.190600] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.190614] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.190792] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.190808] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.191022] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.191037] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.191125] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.191139] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.191255] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.191270] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.191438] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.191453] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.191625] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.191640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.191771] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.191785] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.191948] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.191964] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.192078] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.192092] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.192210] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.192228] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.192356] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.192371] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.192477] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.192491] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.192671] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.192685] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.192873] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.192888] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.193079] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.193094] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.193267] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.193282] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.193391] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.193410] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.670 qpair failed and we were unable to recover it. 00:26:56.670 [2024-07-15 18:12:50.193585] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.670 [2024-07-15 18:12:50.193601] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.193713] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.193727] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.193832] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.193846] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.193956] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.193970] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.194149] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.194165] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.194306] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.194321] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.194434] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.194448] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.194615] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.194630] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.194809] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.194823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.194944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.194958] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.195064] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.195078] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.195267] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.195282] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.195408] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.195423] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.195598] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.195612] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.195722] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.195736] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.195906] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.195922] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.196019] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.196033] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.196202] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.196216] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.196337] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.196353] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.196461] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.196476] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.196578] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.196592] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.196712] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.196726] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.196799] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.196813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.196915] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.196930] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.197103] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.197117] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.197292] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.197307] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.197431] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.197446] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.197555] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.197570] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.197686] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.197701] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.197821] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.197836] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.197937] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.197951] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.198039] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.198053] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.198160] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.198174] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.198276] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.198290] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.198409] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.198423] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.198540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.198554] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.198654] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.198668] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.198836] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.198850] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.198970] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.198984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.199102] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.199119] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.199300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.199315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.199402] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.199416] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.199523] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.199537] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.199635] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.199648] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.199768] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.199782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.199895] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.199909] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.200013] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.200026] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.200205] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.200220] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.200410] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.200425] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.200598] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.200613] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.200792] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.200806] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.200921] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.200935] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.201043] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.201057] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.201156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.201170] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.201286] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.201302] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.201419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.201433] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.201590] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.201604] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.201730] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.201744] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.201862] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.671 [2024-07-15 18:12:50.201877] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.671 qpair failed and we were unable to recover it. 00:26:56.671 [2024-07-15 18:12:50.201991] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.202005] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.202170] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.202184] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.202365] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.202381] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.202505] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.202519] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.202620] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.202634] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.202800] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.202813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.202912] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.202926] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.203035] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.203050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.203218] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.203236] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.203423] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.203438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.203543] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.203558] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.203660] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.203674] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.203782] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.203796] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.203961] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.203979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.204077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.204091] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.204203] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.204228] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.204415] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.204430] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.204620] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.204634] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.204823] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.204837] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.205013] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.205028] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.205203] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.205221] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.205327] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.205341] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.205526] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.205540] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.205660] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.205674] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.205794] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.205808] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.206048] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.206062] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.206194] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.206208] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.206390] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.206405] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.206529] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.206544] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.206702] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.206716] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.206915] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.206928] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.207099] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.207112] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.207238] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.207253] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.207365] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.207379] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.207627] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.207642] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.207809] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.207823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.208080] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.208094] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.208262] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.208276] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.208388] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.208403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.208588] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.208603] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.208784] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.208798] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.208982] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.208997] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.209123] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.209138] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.209223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.209246] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.209424] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.209438] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.209548] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.209562] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.209667] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.209682] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.209800] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.209814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.209986] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.210000] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.210093] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.210107] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.210342] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.210358] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.210456] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.210471] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.210674] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.210688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.210860] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.210874] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.211060] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.211074] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.211195] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.211209] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.211357] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.211372] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.211547] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.211561] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.211741] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.672 [2024-07-15 18:12:50.211755] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.672 qpair failed and we were unable to recover it. 00:26:56.672 [2024-07-15 18:12:50.211871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.211886] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.212053] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.212072] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.212259] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.212274] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.212348] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.212362] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.212476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.212491] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.212657] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.212671] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.212790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.212804] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.213008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.213023] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.213199] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.213213] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.213406] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.213420] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.213541] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.213556] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.213672] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.213686] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.213864] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.213878] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.214123] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.214137] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.214257] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.214272] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.214390] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.214403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.214572] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.214585] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.214760] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.214775] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.214883] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.214898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.215017] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.215032] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.215199] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.215213] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.215288] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.215302] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.215392] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.215406] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.215571] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.215585] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.215743] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.215757] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.215930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.215945] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.216078] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.216093] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.216258] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.216273] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.216485] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.216499] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.216674] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.216689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.216787] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.216801] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.216912] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.216925] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.217035] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.217050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.217235] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.217249] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.217489] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.217503] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.217632] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.217646] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.217747] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.217762] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.217925] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.217939] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.218054] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.218068] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.218152] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.218167] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.218357] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.218371] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.218543] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.218560] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.218670] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.218684] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.218856] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.218871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.218973] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.218987] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.219227] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.219242] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.219411] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.219425] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.219532] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.219545] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.219716] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.219731] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.219890] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.219905] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.220014] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.220029] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.220212] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.220238] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.220413] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.220428] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.220598] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.220612] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.220724] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.220739] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.220917] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.220932] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.221041] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.221055] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.221165] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.221179] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.221305] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.221319] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.221530] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.221544] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.221641] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.673 [2024-07-15 18:12:50.221655] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.673 qpair failed and we were unable to recover it. 00:26:56.673 [2024-07-15 18:12:50.221767] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.221782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.221895] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.221909] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.222093] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.222107] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.222220] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.222239] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.222344] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.222359] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.222460] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.222475] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.222583] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.222598] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.222856] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.222871] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.223040] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.223054] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.223171] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.223185] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.223291] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.223306] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.223439] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.223453] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.223589] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.223603] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.223790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.223804] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.223931] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.223945] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.224056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.224070] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.224256] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.224271] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.224373] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.224387] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.224503] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.224517] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.224624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.224638] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.224736] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.224754] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.224856] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.224870] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.225049] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.225063] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.225161] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.225176] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.225311] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.225325] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.225436] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.225451] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.225619] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.225634] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.225809] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.225823] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.225933] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.225948] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.226056] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.226071] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.226190] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.226206] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.226369] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.226384] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.226621] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.226635] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.226746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.226761] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.226863] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.226878] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.227007] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.227021] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.227172] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.227186] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.227352] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.227366] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.227484] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.227498] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.227672] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.227685] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.227859] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.227873] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.227979] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.227993] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.228120] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.228134] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.228244] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.228258] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.228364] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.228379] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.228469] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.228483] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.228595] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.228609] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.228825] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.228839] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.229036] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.229049] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.229154] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.229167] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.229320] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.229334] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.229445] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.229459] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.229636] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.229649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.229766] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.229780] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.229899] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.229913] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.230013] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.230027] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.230149] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.230163] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.230271] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.230285] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.230521] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.230536] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.230701] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.230715] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.230910] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.230927] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.231038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.674 [2024-07-15 18:12:50.231052] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.674 qpair failed and we were unable to recover it. 00:26:56.674 [2024-07-15 18:12:50.231139] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.231153] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.231388] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.231402] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.231494] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.231508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.231612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.231625] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.231774] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.231788] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.231907] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.231921] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.232036] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.232050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.232228] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.232244] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.232463] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.232476] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.232682] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.232696] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.232773] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.232787] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.232900] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.232914] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.233021] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.233035] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.233248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.233262] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.233386] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.233400] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.233527] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.233540] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.233638] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.233651] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.233766] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.233780] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.233956] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.233970] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.234087] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.234101] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.234210] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.234247] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.234426] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.234440] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.234541] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.234555] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.234629] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.234643] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.234763] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.234776] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.234905] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.234929] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.235051] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.235063] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.235245] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.235256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.235363] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.235373] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.235476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.235487] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.235646] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.235656] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.235766] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.235777] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.235876] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.235886] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.235956] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.235966] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.236063] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.236073] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.236269] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.236280] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.236429] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.236439] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.236600] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.236610] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.236769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.236781] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.236886] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.236897] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.237003] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.237014] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.237182] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.237192] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.237290] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.237301] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.237409] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.237420] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.237520] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.237530] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.237707] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.237717] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.237814] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.237824] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.237945] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.237955] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.238048] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.238058] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.238164] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.238174] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.238248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.238258] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.238424] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.238435] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.238533] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.238543] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.238677] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.238688] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.675 [2024-07-15 18:12:50.238794] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.675 [2024-07-15 18:12:50.238804] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.675 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.238886] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.238895] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.238994] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.239004] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.239107] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.239117] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.239223] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.239237] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.239550] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.239561] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.239728] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.239738] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.239834] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.239844] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.240016] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.240027] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.240127] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.240137] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.240317] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.240327] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.240425] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.240435] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.240529] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.240540] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.240637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.240647] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.240748] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.240758] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.240873] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.240883] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.241033] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.241043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.241141] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.241151] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.241253] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.241263] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.241356] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.241366] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.241468] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.241479] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.241652] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.241662] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.241832] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.241841] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.241940] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.241950] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.242043] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.242054] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.242150] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.242160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.242271] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.242281] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.242393] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.242403] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.242515] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.242525] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.242714] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.242725] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.242806] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.242816] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.242913] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.242923] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.243038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.243049] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.243152] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.243162] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.243322] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.243332] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.243498] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.243508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.243614] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.243624] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.243721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.243731] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.243824] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.243835] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.244010] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.244020] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.244194] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.244203] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.244311] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.244322] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.244563] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.244574] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.244694] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.244705] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.244796] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.244806] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.244907] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.244917] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.245038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.245048] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.245209] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.245219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.245336] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.245347] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.245586] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.245596] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.245700] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.245711] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.245883] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.245899] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.246004] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.246018] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.246119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.246133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.246301] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.246315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.246495] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.246508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.246691] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.246705] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.246942] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.246955] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.247072] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.247086] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.247187] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.247200] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.247436] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.247451] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.247624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.247638] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.247843] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.247858] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.247965] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.247979] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.248083] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.248100] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.248286] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.248300] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.248474] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.248487] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.248652] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.248665] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.248833] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.248846] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.249088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.249101] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.249219] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.249239] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.249428] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.249442] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.249619] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.249632] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.249916] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.249930] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.676 [2024-07-15 18:12:50.250164] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.676 [2024-07-15 18:12:50.250177] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.676 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.250307] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.250321] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.250434] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.250448] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.250612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.250626] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.250800] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.250813] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.250931] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.250944] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.251060] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.251074] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.251203] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.251218] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.251403] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.251417] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.251516] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.251530] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.251647] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.251661] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.251794] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.251807] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.251994] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.252008] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.252160] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.252174] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:26:56.677 [2024-07-15 18:12:50.252363] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.252377] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.252550] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.252564] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@862 -- # return 0 00:26:56.677 [2024-07-15 18:12:50.252668] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.252684] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.252784] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.252799] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.252914] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.252928] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:26:56.677 [2024-07-15 18:12:50.253028] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.253042] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.253162] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.253176] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@728 -- # xtrace_disable 00:26:56.677 [2024-07-15 18:12:50.253364] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.253379] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.253499] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.253513] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:56.677 [2024-07-15 18:12:50.253638] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.253652] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.253800] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.253814] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.254070] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.254084] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.254240] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.254254] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.254517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.254532] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.254736] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.254751] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.254865] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.254879] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.254980] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.254994] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.255102] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.255117] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.255278] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.255291] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.255456] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.255470] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.255645] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.255659] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.255766] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.255780] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.255884] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.255898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.256150] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.256164] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.256358] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.256375] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.256554] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.256569] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.256814] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.256828] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.256943] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.256958] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.257148] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.257164] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.257288] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.257303] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.257482] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.257500] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.257631] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.257647] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.257759] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.257773] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.257950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.257963] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.258131] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.258145] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.258236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.258251] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.258356] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.258370] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.258467] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.258480] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.258646] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.258660] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.258859] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.258873] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.259043] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.259056] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.259181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.259198] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.259324] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.259340] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.259445] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.259459] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.259637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.259650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.259775] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.259789] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.259916] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.259930] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.260082] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.260098] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.260204] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.260218] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.260301] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.260315] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.260438] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.677 [2024-07-15 18:12:50.260451] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.677 qpair failed and we were unable to recover it. 00:26:56.677 [2024-07-15 18:12:50.260631] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.260645] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.260830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.260844] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.261021] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.261035] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.261156] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.261171] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.261289] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.261304] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.261409] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.261424] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.261660] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.261674] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.261874] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.261888] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.262011] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.262026] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.262149] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.262163] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.262281] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.262296] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.262418] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.262432] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.262543] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.262557] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.262696] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.262709] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.262888] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.262901] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.263009] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.263023] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.263188] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.263203] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.263361] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.263377] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.263484] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.263498] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.263705] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.263718] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.263825] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.263838] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.263939] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.263953] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.264037] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.264050] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.264158] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.264172] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.264294] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.264310] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.264419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.264434] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.264544] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.264559] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.264675] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.264689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.264789] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.264804] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.264978] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.264993] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.265107] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.265125] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.265244] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.265260] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.265365] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.265379] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.265493] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.265507] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.265679] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.265693] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.265812] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.265826] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.265937] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.265951] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.266052] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.266066] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.266172] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.266186] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.266298] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.266312] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.266426] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.266441] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.266543] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.266557] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.266676] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.266689] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.266883] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.266898] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.267021] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.267036] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.267139] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.267152] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.267251] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.267265] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.267371] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.267385] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.267502] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.267516] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.267615] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.267630] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.267773] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.267786] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.267897] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.267911] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.268020] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.268033] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.268199] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.268213] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.268326] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.268339] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.268443] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.268456] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.268572] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.268586] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.268685] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.268698] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.268813] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.268827] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.268947] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.268961] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c0000b90 with addr=10.0.0.2, port=4420 00:26:56.678 qpair failed and we were unable to recover it. 00:26:56.678 [2024-07-15 18:12:50.269127] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.678 [2024-07-15 18:12:50.269139] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.269220] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.269236] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.269341] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.269352] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.269509] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.269519] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.269630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.269640] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.269729] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.269739] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.269847] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.269857] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.269953] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.269963] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.270067] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.270077] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.270173] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.270183] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.270281] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.270294] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.270379] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.270389] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.270508] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.270518] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.270610] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.270620] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.270745] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.270756] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.270857] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.270868] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.270964] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.270974] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.271131] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.271141] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.271242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.271253] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.271347] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.271358] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.271457] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.271467] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.271572] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.271582] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.271676] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.271686] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.271785] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.271795] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.271892] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.271901] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.272061] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.272072] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.272215] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.272229] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.272326] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.272338] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.272447] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.272457] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.272558] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.272570] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.272669] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.272679] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.272815] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.272825] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.272922] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.272932] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.273025] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.273034] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.273126] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.273136] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.273246] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.273257] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.273349] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.273359] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.273523] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.273534] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.273642] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.273652] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.273749] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.273761] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.273899] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.273909] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.274008] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.274017] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.274123] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.274134] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.274246] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.274256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.274350] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.274360] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.274520] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.274531] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.274621] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.274631] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.274791] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.274800] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.274917] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.274927] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.275027] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.275037] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.275127] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.275139] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.275248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.275258] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.275364] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.275375] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.275472] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.275482] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.275580] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.275590] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.275701] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.275711] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.275823] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.275833] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.679 [2024-07-15 18:12:50.275931] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.679 [2024-07-15 18:12:50.275942] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.679 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.276036] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.276046] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.276152] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.276162] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.276280] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.276290] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.276399] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.276409] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.276503] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.276513] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.276613] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.276623] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.276737] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.276747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.276856] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.276866] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.277016] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.277026] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.277143] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.277154] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.277258] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.277269] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.277362] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.277374] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.277466] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.277476] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.277569] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.277579] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.277671] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.277681] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.277782] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.277792] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.277889] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.277899] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.277986] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.277996] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.278105] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.278116] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.278220] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.278244] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.278352] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.278363] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.278433] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.278443] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.278582] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.278593] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.278683] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.278693] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.278823] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.278832] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.278920] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.278930] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.279088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.279098] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.279193] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.279203] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.279296] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.279307] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.279408] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.279418] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.279530] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.279540] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.279635] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.279647] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.279751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.279763] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.279880] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.279890] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.280054] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.280065] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.280168] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.280178] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.280282] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.280292] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.280432] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.280444] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.280538] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.280548] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.280642] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.280652] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.280748] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.280758] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.280872] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.280882] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.280975] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.280985] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.281149] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.281159] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.281277] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.281287] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.281397] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.281408] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.281513] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.281523] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.281627] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.281637] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.281728] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.281738] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.281841] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.281851] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.281947] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.281957] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.282055] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.282064] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.282157] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.282167] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.282267] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.282277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.282382] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.282392] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.282501] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.282511] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.282615] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.282624] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.282786] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.282797] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.282895] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.282905] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.283002] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.283012] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.283164] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.283174] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.283291] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.283302] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.283436] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.283446] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.283545] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.283557] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.680 [2024-07-15 18:12:50.283658] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.680 [2024-07-15 18:12:50.283668] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.680 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.283756] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.283766] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.283869] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.283880] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.283973] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.283983] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.284139] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.284149] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.284222] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.284236] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.284348] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.284359] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.284492] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.284504] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.284604] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.284618] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.284712] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.284722] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.284823] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.284833] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.284926] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.284937] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.285032] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.285043] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.285150] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.285160] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.285270] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.285280] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.285387] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.285397] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.285493] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.285503] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.285602] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.285612] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.285704] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.285714] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.285805] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.285815] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.285929] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.285940] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.286049] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.286061] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.286133] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.286145] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.286263] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.286273] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.286367] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.286378] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.286479] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.286489] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.286584] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.286594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.286690] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.286700] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.286798] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.286808] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.286909] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.286920] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.287030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.287040] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.287139] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.681 [2024-07-15 18:12:50.287148] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.681 qpair failed and we were unable to recover it. 00:26:56.681 [2024-07-15 18:12:50.287253] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.287264] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.287373] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.287383] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.287479] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.287489] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.287612] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.287633] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x1931ed0 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.287754] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.287776] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.287862] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.287876] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.287978] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.287992] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.288096] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.288111] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:26:56.682 [2024-07-15 18:12:50.288218] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.288239] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.288385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.288398] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.288525] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.288540] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@19 -- # rpc_cmd bdev_malloc_create 64 512 -b Malloc0 00:26:56.682 [2024-07-15 18:12:50.288654] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.288669] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.288776] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.288790] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:56.682 [2024-07-15 18:12:50.288896] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.288912] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.288988] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.289001] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8d0000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.289168] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.289182] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:56.682 [2024-07-15 18:12:50.289280] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.289292] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.289383] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.289394] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.289494] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.289504] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.289607] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.289617] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.289716] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.289726] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.289833] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.289843] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.289944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.289954] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.290062] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.290073] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.290168] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.290178] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.290287] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.290298] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.290408] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.290418] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.290512] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.290523] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.290619] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.290629] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.290728] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.290738] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.290830] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.290840] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.290933] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.290943] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.291046] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.291056] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.291150] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.291159] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.291324] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.291336] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.291446] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.291457] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.291555] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.291565] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.291659] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.682 [2024-07-15 18:12:50.291669] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.682 qpair failed and we were unable to recover it. 00:26:56.682 [2024-07-15 18:12:50.291776] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.291786] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.291877] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.291887] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.291986] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.291996] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.292088] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.292100] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.292208] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.292219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.292339] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.292349] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.292443] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.292453] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.292546] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.292556] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.292650] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.292660] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.292758] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.292768] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.292860] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.292870] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.292944] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.292955] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.293053] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.293063] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.293155] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.293165] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.293276] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.293287] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.293361] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.293371] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.293476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.293486] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.293580] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.293591] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.293681] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.293693] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.293790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.293801] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.293896] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.293906] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.294012] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.294022] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.294119] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.294129] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.294228] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.294239] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.294386] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.294397] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.294532] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.294542] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.294636] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.294646] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.294744] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.294754] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.294841] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.294851] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.294946] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.294956] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.295049] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.295059] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.295153] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.295163] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.295263] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.295274] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.295385] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.295395] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.295557] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.295567] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.295662] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.295672] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.295769] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.295779] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.295870] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.295880] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.295975] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.295984] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.296079] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.683 [2024-07-15 18:12:50.296089] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.683 qpair failed and we were unable to recover it. 00:26:56.683 [2024-07-15 18:12:50.296185] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.296194] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.296300] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.296310] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.296513] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.296523] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.296628] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.296639] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.296738] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.296747] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.296845] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.296854] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.296959] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.296969] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.297111] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.297121] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.297220] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.297240] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.297396] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.297407] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.297637] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.297647] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.297745] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.297755] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.297848] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.297858] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.297955] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.297965] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.298085] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.298096] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.298193] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.298203] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.298298] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.298308] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.298409] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.298419] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.298517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.298527] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.298644] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.298654] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.298746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.298756] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.298851] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.298861] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.298955] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.298964] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.299054] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.299064] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.299237] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.299247] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.299345] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.299355] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.299520] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.299530] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.299624] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.299634] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.299802] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.299812] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.299905] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.299915] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.300077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.300087] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.300183] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.300194] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.300292] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.300303] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.300400] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.300410] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.300508] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.300518] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.300682] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.300692] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.300793] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.300803] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.300915] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.300925] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.301020] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.684 [2024-07-15 18:12:50.301030] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.684 qpair failed and we were unable to recover it. 00:26:56.684 [2024-07-15 18:12:50.301140] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.301151] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.301242] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.301253] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.301404] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.301415] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.301531] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.301541] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.301638] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.301650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.301752] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.301762] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.301871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.301880] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.302037] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.302049] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.302249] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.302260] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.302422] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.302432] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.302546] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.302557] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.302786] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.302797] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.302977] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.302989] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.303060] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.303071] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.303245] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.303256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.303423] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.303434] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.303638] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.303650] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.303759] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.303770] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.303978] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.303989] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.304159] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.304170] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.304257] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.304270] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.304381] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.304392] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.304560] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.304572] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.304804] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.304816] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.304929] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.304940] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.305172] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.305185] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.305295] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.305305] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.305477] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.305489] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.305603] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.305614] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.305726] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.305738] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.305861] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.305873] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.306054] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.306065] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.306246] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.306257] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.306476] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.685 [2024-07-15 18:12:50.306487] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.685 qpair failed and we were unable to recover it. 00:26:56.685 [2024-07-15 18:12:50.306588] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.306598] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.306761] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.306772] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.306964] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.306974] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.307072] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.307082] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.307248] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.307259] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.307509] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.307520] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.307718] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.307728] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.307832] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.307842] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.307938] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.307948] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.308107] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.308118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.308301] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.308312] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.308443] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.308453] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 Malloc0 00:26:56.686 [2024-07-15 18:12:50.308545] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.308555] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.308670] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.308679] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.308784] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.308794] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.308914] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.308924] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.309107] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.309117] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:56.686 [2024-07-15 18:12:50.309214] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.309228] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.309337] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.309346] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.309450] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.309459] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.309567] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.309577] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@21 -- # rpc_cmd nvmf_create_transport -t tcp -o 00:26:56.686 [2024-07-15 18:12:50.309764] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.309774] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.309871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.309881] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:56.686 [2024-07-15 18:12:50.310000] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.310010] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.310115] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.310125] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:56.686 [2024-07-15 18:12:50.310236] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.310246] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.310408] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.310418] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.310527] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.310536] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.310662] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.310672] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.310742] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.310751] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.310938] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.310947] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.311054] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.311064] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.311165] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.311175] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.311286] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.311296] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.311395] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.311405] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.311518] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.311530] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.311727] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.311736] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.311858] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.311868] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.686 [2024-07-15 18:12:50.312067] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.686 [2024-07-15 18:12:50.312076] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.686 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.312176] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.312186] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.312287] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.312297] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.312460] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.312469] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.312586] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.312595] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.312696] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.312706] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.312934] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.312944] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.313103] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.313113] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.313210] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.313219] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.313348] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.313358] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.313454] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.313463] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.313690] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.313700] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.313813] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.313822] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.313993] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.314003] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.314233] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.314243] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.314400] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.314410] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.314525] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.314535] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.314716] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.314725] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.314900] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.314909] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.315071] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.315080] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.315267] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.315277] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.315461] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.315470] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.315581] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.315590] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.315823] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.315832] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.315994] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.316004] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.316108] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.316118] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.316154] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:26:56.687 [2024-07-15 18:12:50.316342] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.316352] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.316534] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.316544] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.316772] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.316782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.317013] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.317023] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.317134] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.317143] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.317320] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.317330] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.317441] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.317451] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.317626] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.317636] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.317736] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.317746] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.317930] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.317940] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.318100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.318110] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.318289] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.318300] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.318473] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.318483] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.318722] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.687 [2024-07-15 18:12:50.318731] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.687 qpair failed and we were unable to recover it. 00:26:56.687 [2024-07-15 18:12:50.318856] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.318866] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.318968] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.318977] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.319149] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.319158] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.319276] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.319286] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.319458] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.319468] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.319648] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.319658] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.319754] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.319763] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.319926] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.319935] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.320124] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.320133] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.320250] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.320259] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.320443] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.320455] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.320569] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.320578] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.320687] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.320697] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.320823] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.320833] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.320934] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.320943] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.321041] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.321051] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.321211] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.321221] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.321424] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.321434] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.321610] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.321621] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:56.688 [2024-07-15 18:12:50.321790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.321800] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.321910] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.321920] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 00:26:56.688 [2024-07-15 18:12:50.322083] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.322093] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:56.688 [2024-07-15 18:12:50.322325] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.322338] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:56.688 [2024-07-15 18:12:50.322592] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.322602] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.322762] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.322771] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.322950] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.322959] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.323141] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.323150] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.323326] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.323336] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.323512] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.323522] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.323752] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.323761] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.323923] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.323932] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.324110] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.324120] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.324238] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.324248] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.324477] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.324487] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.324663] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.324672] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.324834] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.324846] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.325077] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.325087] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.325275] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.325285] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.688 [2024-07-15 18:12:50.325407] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.688 [2024-07-15 18:12:50.325417] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.688 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.325541] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.325551] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.325721] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.325730] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.325894] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.325904] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.326009] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.326018] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.326181] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.326190] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.326296] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.326307] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.326516] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.326526] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.326763] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.326773] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.326871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.326880] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.326993] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.327002] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.327111] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.327120] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.327270] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.327280] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.327517] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.327527] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.327630] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.327639] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.327874] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.327884] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.328042] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.328051] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.328246] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.328256] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.328419] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.328429] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.328539] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.328549] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.328751] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.328761] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.328932] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.328942] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.329064] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.329074] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.329265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.329275] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.329528] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.329539] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.329746] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.329755] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.329923] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.329933] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.330100] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.330109] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.330285] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.330295] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.330403] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.330412] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.330521] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.330530] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.330704] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.330714] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.330965] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.330974] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.331082] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.331091] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.331191] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.331200] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.689 qpair failed and we were unable to recover it. 00:26:56.689 [2024-07-15 18:12:50.331314] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.689 [2024-07-15 18:12:50.331324] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.331499] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.331508] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.331604] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.331615] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.331871] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.331880] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.332113] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.332123] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.332214] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.332227] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.332342] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.332352] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.332551] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.332561] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.332742] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.332752] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.332872] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.332881] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.333117] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.333126] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.333301] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.333311] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.333523] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.333532] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:56.690 [2024-07-15 18:12:50.333689] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.333699] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.333810] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.333819] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@24 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Malloc0 00:26:56.690 [2024-07-15 18:12:50.334050] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.334061] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.334289] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.334300] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.334462] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.334472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:56.690 [2024-07-15 18:12:50.334631] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.334642] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.334760] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.334769] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.334996] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.335006] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.335126] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.335136] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.335367] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.335376] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.335496] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.335505] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.335609] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.335618] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.335845] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.335854] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.336038] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.336048] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.336234] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.336246] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.336367] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.336377] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.336551] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.336561] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.336758] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.336767] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.336880] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.336890] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.337000] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.337010] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.337182] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.337191] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.337412] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.337421] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.337531] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.337541] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.337639] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.337649] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.690 [2024-07-15 18:12:50.337772] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.690 [2024-07-15 18:12:50.337782] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.690 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.337960] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.337970] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.338129] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.338139] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.338265] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.338275] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.338386] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.338396] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.338574] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.338583] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.338790] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.338800] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.338914] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.338924] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.339086] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.339096] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.339259] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.339268] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.339381] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.339391] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.339584] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.339594] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.339753] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.339762] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.339885] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.339895] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.339968] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.339978] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.340251] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.340261] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.340462] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.340472] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.340648] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.340657] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.340771] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.340780] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.340895] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.340905] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.340999] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.341008] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.341116] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.341126] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.341244] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.341254] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.341365] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.341375] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.341485] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.341494] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:56.691 [2024-07-15 18:12:50.341666] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.341683] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.341859] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.341869] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.342030] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@25 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:26:56.691 [2024-07-15 18:12:50.342040] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.342270] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.342280] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:56.691 [2024-07-15 18:12:50.342389] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.342399] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.342578] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.342588] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:56.691 [2024-07-15 18:12:50.342758] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.342768] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.342938] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.342947] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.343200] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.343210] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.343309] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.343318] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.343439] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.343449] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.343540] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.343550] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.343757] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.343766] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.691 [2024-07-15 18:12:50.343869] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.691 [2024-07-15 18:12:50.343879] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.691 qpair failed and we were unable to recover it. 00:26:56.692 [2024-07-15 18:12:50.343972] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.692 [2024-07-15 18:12:50.343982] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.692 qpair failed and we were unable to recover it. 00:26:56.692 [2024-07-15 18:12:50.344216] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.692 [2024-07-15 18:12:50.344228] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.692 qpair failed and we were unable to recover it. 00:26:56.692 [2024-07-15 18:12:50.344332] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.692 [2024-07-15 18:12:50.344341] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.692 qpair failed and we were unable to recover it. 00:26:56.692 [2024-07-15 18:12:50.344455] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.692 [2024-07-15 18:12:50.344465] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.692 qpair failed and we were unable to recover it. 00:26:56.692 [2024-07-15 18:12:50.344641] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.692 [2024-07-15 18:12:50.344651] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.692 qpair failed and we were unable to recover it. 00:26:56.692 [2024-07-15 18:12:50.344810] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.692 [2024-07-15 18:12:50.344819] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.692 qpair failed and we were unable to recover it. 00:26:56.692 [2024-07-15 18:12:50.344985] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.692 [2024-07-15 18:12:50.344995] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.692 qpair failed and we were unable to recover it. 00:26:56.692 [2024-07-15 18:12:50.345106] posix.c:1038:posix_sock_create: *ERROR*: connect() failed, errno = 111 00:26:56.692 [2024-07-15 18:12:50.345116] nvme_tcp.c:2383:nvme_tcp_qpair_connect_sock: *ERROR*: sock connection error of tqpair=0x7fc8c8000b90 with addr=10.0.0.2, port=4420 00:26:56.692 qpair failed and we were unable to recover it. 00:26:56.692 [2024-07-15 18:12:50.345153] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:26:56.692 [2024-07-15 18:12:50.346652] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.692 [2024-07-15 18:12:50.346734] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.692 [2024-07-15 18:12:50.346756] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.692 [2024-07-15 18:12:50.346763] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.692 [2024-07-15 18:12:50.346770] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.692 [2024-07-15 18:12:50.346790] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.692 qpair failed and we were unable to recover it. 00:26:56.692 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:56.692 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@26 -- # rpc_cmd nvmf_subsystem_add_listener discovery -t tcp -a 10.0.0.2 -s 4420 00:26:56.692 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@559 -- # xtrace_disable 00:26:56.692 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:26:56.952 [2024-07-15 18:12:50.356677] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.952 [2024-07-15 18:12:50.356743] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.952 [2024-07-15 18:12:50.356761] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.952 [2024-07-15 18:12:50.356768] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.952 [2024-07-15 18:12:50.356774] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.952 [2024-07-15 18:12:50.356790] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.952 qpair failed and we were unable to recover it. 00:26:56.952 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:26:56.952 18:12:50 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@50 -- # wait 746092 00:26:56.952 [2024-07-15 18:12:50.366693] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.952 [2024-07-15 18:12:50.366776] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.952 [2024-07-15 18:12:50.366791] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.952 [2024-07-15 18:12:50.366798] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.952 [2024-07-15 18:12:50.366804] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.952 [2024-07-15 18:12:50.366820] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.952 qpair failed and we were unable to recover it. 00:26:56.952 [2024-07-15 18:12:50.376625] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.952 [2024-07-15 18:12:50.376696] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.952 [2024-07-15 18:12:50.376711] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.952 [2024-07-15 18:12:50.376718] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.952 [2024-07-15 18:12:50.376724] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.952 [2024-07-15 18:12:50.376738] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.952 qpair failed and we were unable to recover it. 00:26:56.952 [2024-07-15 18:12:50.386706] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.952 [2024-07-15 18:12:50.386773] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.952 [2024-07-15 18:12:50.386791] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.952 [2024-07-15 18:12:50.386798] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.952 [2024-07-15 18:12:50.386805] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.952 [2024-07-15 18:12:50.386821] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.952 qpair failed and we were unable to recover it. 00:26:56.952 [2024-07-15 18:12:50.396715] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.952 [2024-07-15 18:12:50.396825] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.952 [2024-07-15 18:12:50.396840] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.952 [2024-07-15 18:12:50.396847] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.952 [2024-07-15 18:12:50.396854] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.952 [2024-07-15 18:12:50.396870] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.952 qpair failed and we were unable to recover it. 00:26:56.952 [2024-07-15 18:12:50.406732] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.953 [2024-07-15 18:12:50.406840] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.953 [2024-07-15 18:12:50.406859] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.953 [2024-07-15 18:12:50.406866] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.953 [2024-07-15 18:12:50.406873] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.953 [2024-07-15 18:12:50.406889] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.953 qpair failed and we were unable to recover it. 00:26:56.953 [2024-07-15 18:12:50.416725] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.953 [2024-07-15 18:12:50.416790] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.953 [2024-07-15 18:12:50.416806] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.953 [2024-07-15 18:12:50.416814] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.953 [2024-07-15 18:12:50.416819] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.953 [2024-07-15 18:12:50.416835] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.953 qpair failed and we were unable to recover it. 00:26:56.953 [2024-07-15 18:12:50.426778] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.953 [2024-07-15 18:12:50.426893] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.953 [2024-07-15 18:12:50.426908] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.953 [2024-07-15 18:12:50.426915] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.953 [2024-07-15 18:12:50.426921] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.953 [2024-07-15 18:12:50.426935] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.953 qpair failed and we were unable to recover it. 00:26:56.953 [2024-07-15 18:12:50.436774] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.953 [2024-07-15 18:12:50.436863] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.953 [2024-07-15 18:12:50.436880] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.953 [2024-07-15 18:12:50.436888] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.953 [2024-07-15 18:12:50.436894] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.953 [2024-07-15 18:12:50.436909] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.953 qpair failed and we were unable to recover it. 00:26:56.953 [2024-07-15 18:12:50.446846] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.953 [2024-07-15 18:12:50.446910] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.953 [2024-07-15 18:12:50.446925] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.953 [2024-07-15 18:12:50.446932] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.953 [2024-07-15 18:12:50.446941] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.953 [2024-07-15 18:12:50.446956] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.953 qpair failed and we were unable to recover it. 00:26:56.953 [2024-07-15 18:12:50.456833] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.953 [2024-07-15 18:12:50.456898] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.953 [2024-07-15 18:12:50.456914] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.953 [2024-07-15 18:12:50.456921] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.953 [2024-07-15 18:12:50.456927] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.953 [2024-07-15 18:12:50.456942] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.953 qpair failed and we were unable to recover it. 00:26:56.953 [2024-07-15 18:12:50.466884] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.953 [2024-07-15 18:12:50.466949] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.953 [2024-07-15 18:12:50.466965] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.953 [2024-07-15 18:12:50.466972] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.953 [2024-07-15 18:12:50.466978] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.953 [2024-07-15 18:12:50.466994] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.953 qpair failed and we were unable to recover it. 00:26:56.953 [2024-07-15 18:12:50.476905] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.953 [2024-07-15 18:12:50.476968] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.953 [2024-07-15 18:12:50.476984] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.953 [2024-07-15 18:12:50.476990] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.953 [2024-07-15 18:12:50.476996] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.953 [2024-07-15 18:12:50.477011] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.953 qpair failed and we were unable to recover it. 00:26:56.953 [2024-07-15 18:12:50.486920] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.953 [2024-07-15 18:12:50.486985] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.953 [2024-07-15 18:12:50.487000] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.953 [2024-07-15 18:12:50.487007] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.953 [2024-07-15 18:12:50.487012] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.953 [2024-07-15 18:12:50.487027] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.953 qpair failed and we were unable to recover it. 00:26:56.953 [2024-07-15 18:12:50.496969] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.953 [2024-07-15 18:12:50.497041] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.953 [2024-07-15 18:12:50.497056] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.953 [2024-07-15 18:12:50.497063] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.953 [2024-07-15 18:12:50.497069] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.953 [2024-07-15 18:12:50.497084] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.953 qpair failed and we were unable to recover it. 00:26:56.953 [2024-07-15 18:12:50.507010] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.953 [2024-07-15 18:12:50.507077] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.953 [2024-07-15 18:12:50.507091] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.953 [2024-07-15 18:12:50.507099] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.953 [2024-07-15 18:12:50.507106] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.953 [2024-07-15 18:12:50.507121] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.953 qpair failed and we were unable to recover it. 00:26:56.953 [2024-07-15 18:12:50.517021] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.953 [2024-07-15 18:12:50.517085] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.953 [2024-07-15 18:12:50.517100] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.953 [2024-07-15 18:12:50.517108] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.953 [2024-07-15 18:12:50.517114] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.953 [2024-07-15 18:12:50.517128] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.953 qpair failed and we were unable to recover it. 00:26:56.953 [2024-07-15 18:12:50.527070] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.953 [2024-07-15 18:12:50.527151] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.953 [2024-07-15 18:12:50.527167] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.953 [2024-07-15 18:12:50.527175] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.953 [2024-07-15 18:12:50.527182] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.953 [2024-07-15 18:12:50.527197] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.954 qpair failed and we were unable to recover it. 00:26:56.954 [2024-07-15 18:12:50.537080] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.954 [2024-07-15 18:12:50.537145] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.954 [2024-07-15 18:12:50.537161] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.954 [2024-07-15 18:12:50.537170] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.954 [2024-07-15 18:12:50.537180] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.954 [2024-07-15 18:12:50.537195] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.954 qpair failed and we were unable to recover it. 00:26:56.954 [2024-07-15 18:12:50.547085] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.954 [2024-07-15 18:12:50.547147] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.954 [2024-07-15 18:12:50.547161] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.954 [2024-07-15 18:12:50.547168] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.954 [2024-07-15 18:12:50.547174] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.954 [2024-07-15 18:12:50.547189] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.954 qpair failed and we were unable to recover it. 00:26:56.954 [2024-07-15 18:12:50.557131] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.954 [2024-07-15 18:12:50.557190] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.954 [2024-07-15 18:12:50.557204] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.954 [2024-07-15 18:12:50.557211] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.954 [2024-07-15 18:12:50.557218] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.954 [2024-07-15 18:12:50.557237] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.954 qpair failed and we were unable to recover it. 00:26:56.954 [2024-07-15 18:12:50.567232] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.954 [2024-07-15 18:12:50.567322] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.954 [2024-07-15 18:12:50.567337] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.954 [2024-07-15 18:12:50.567345] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.954 [2024-07-15 18:12:50.567351] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.954 [2024-07-15 18:12:50.567366] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.954 qpair failed and we were unable to recover it. 00:26:56.954 [2024-07-15 18:12:50.577263] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.954 [2024-07-15 18:12:50.577330] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.954 [2024-07-15 18:12:50.577346] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.954 [2024-07-15 18:12:50.577353] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.954 [2024-07-15 18:12:50.577360] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.954 [2024-07-15 18:12:50.577376] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.954 qpair failed and we were unable to recover it. 00:26:56.954 [2024-07-15 18:12:50.587277] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.954 [2024-07-15 18:12:50.587346] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.954 [2024-07-15 18:12:50.587361] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.954 [2024-07-15 18:12:50.587369] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.954 [2024-07-15 18:12:50.587375] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.954 [2024-07-15 18:12:50.587390] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.954 qpair failed and we were unable to recover it. 00:26:56.954 [2024-07-15 18:12:50.597289] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.954 [2024-07-15 18:12:50.597351] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.954 [2024-07-15 18:12:50.597365] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.954 [2024-07-15 18:12:50.597372] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.954 [2024-07-15 18:12:50.597378] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.954 [2024-07-15 18:12:50.597393] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.954 qpair failed and we were unable to recover it. 00:26:56.954 [2024-07-15 18:12:50.607391] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.954 [2024-07-15 18:12:50.607463] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.954 [2024-07-15 18:12:50.607478] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.954 [2024-07-15 18:12:50.607485] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.954 [2024-07-15 18:12:50.607491] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.954 [2024-07-15 18:12:50.607506] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.954 qpair failed and we were unable to recover it. 00:26:56.954 [2024-07-15 18:12:50.617299] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.954 [2024-07-15 18:12:50.617375] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.954 [2024-07-15 18:12:50.617390] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.954 [2024-07-15 18:12:50.617397] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.954 [2024-07-15 18:12:50.617403] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.954 [2024-07-15 18:12:50.617417] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.954 qpair failed and we were unable to recover it. 00:26:56.954 [2024-07-15 18:12:50.627345] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.954 [2024-07-15 18:12:50.627408] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.954 [2024-07-15 18:12:50.627423] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.954 [2024-07-15 18:12:50.627433] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.954 [2024-07-15 18:12:50.627439] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.954 [2024-07-15 18:12:50.627454] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.954 qpair failed and we were unable to recover it. 00:26:56.954 [2024-07-15 18:12:50.637367] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.954 [2024-07-15 18:12:50.637428] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.954 [2024-07-15 18:12:50.637442] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.954 [2024-07-15 18:12:50.637449] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.954 [2024-07-15 18:12:50.637456] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.954 [2024-07-15 18:12:50.637471] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.954 qpair failed and we were unable to recover it. 00:26:56.954 [2024-07-15 18:12:50.647398] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.954 [2024-07-15 18:12:50.647490] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.954 [2024-07-15 18:12:50.647505] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.954 [2024-07-15 18:12:50.647512] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.954 [2024-07-15 18:12:50.647518] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.954 [2024-07-15 18:12:50.647533] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.954 qpair failed and we were unable to recover it. 00:26:56.954 [2024-07-15 18:12:50.657395] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.954 [2024-07-15 18:12:50.657480] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.954 [2024-07-15 18:12:50.657495] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.954 [2024-07-15 18:12:50.657503] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.954 [2024-07-15 18:12:50.657509] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.954 [2024-07-15 18:12:50.657523] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.954 qpair failed and we were unable to recover it. 00:26:56.954 [2024-07-15 18:12:50.667368] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.954 [2024-07-15 18:12:50.667430] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.954 [2024-07-15 18:12:50.667445] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.954 [2024-07-15 18:12:50.667452] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.954 [2024-07-15 18:12:50.667458] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.954 [2024-07-15 18:12:50.667473] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.954 qpair failed and we were unable to recover it. 00:26:56.954 [2024-07-15 18:12:50.677438] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:56.954 [2024-07-15 18:12:50.677524] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:56.955 [2024-07-15 18:12:50.677539] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:56.955 [2024-07-15 18:12:50.677546] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:56.955 [2024-07-15 18:12:50.677552] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:56.955 [2024-07-15 18:12:50.677566] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:56.955 qpair failed and we were unable to recover it. 00:26:57.214 [2024-07-15 18:12:50.687427] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.214 [2024-07-15 18:12:50.687522] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.214 [2024-07-15 18:12:50.687537] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.214 [2024-07-15 18:12:50.687544] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.214 [2024-07-15 18:12:50.687550] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.215 [2024-07-15 18:12:50.687565] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.215 qpair failed and we were unable to recover it. 00:26:57.215 [2024-07-15 18:12:50.697534] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.215 [2024-07-15 18:12:50.697600] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.215 [2024-07-15 18:12:50.697616] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.215 [2024-07-15 18:12:50.697623] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.215 [2024-07-15 18:12:50.697629] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.215 [2024-07-15 18:12:50.697644] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.215 qpair failed and we were unable to recover it. 00:26:57.215 [2024-07-15 18:12:50.707561] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.215 [2024-07-15 18:12:50.707635] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.215 [2024-07-15 18:12:50.707650] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.215 [2024-07-15 18:12:50.707657] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.215 [2024-07-15 18:12:50.707663] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.215 [2024-07-15 18:12:50.707678] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.215 qpair failed and we were unable to recover it. 00:26:57.215 [2024-07-15 18:12:50.717574] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.215 [2024-07-15 18:12:50.717636] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.215 [2024-07-15 18:12:50.717654] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.215 [2024-07-15 18:12:50.717661] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.215 [2024-07-15 18:12:50.717667] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.215 [2024-07-15 18:12:50.717682] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.215 qpair failed and we were unable to recover it. 00:26:57.215 [2024-07-15 18:12:50.727572] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.215 [2024-07-15 18:12:50.727645] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.215 [2024-07-15 18:12:50.727660] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.215 [2024-07-15 18:12:50.727667] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.215 [2024-07-15 18:12:50.727673] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.215 [2024-07-15 18:12:50.727689] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.215 qpair failed and we were unable to recover it. 00:26:57.215 [2024-07-15 18:12:50.737627] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.215 [2024-07-15 18:12:50.737700] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.215 [2024-07-15 18:12:50.737715] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.215 [2024-07-15 18:12:50.737722] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.215 [2024-07-15 18:12:50.737728] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.215 [2024-07-15 18:12:50.737743] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.215 qpair failed and we were unable to recover it. 00:26:57.215 [2024-07-15 18:12:50.747677] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.215 [2024-07-15 18:12:50.747787] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.215 [2024-07-15 18:12:50.747802] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.215 [2024-07-15 18:12:50.747809] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.215 [2024-07-15 18:12:50.747816] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.215 [2024-07-15 18:12:50.747831] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.215 qpair failed and we were unable to recover it. 00:26:57.215 [2024-07-15 18:12:50.757738] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.215 [2024-07-15 18:12:50.757802] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.215 [2024-07-15 18:12:50.757816] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.215 [2024-07-15 18:12:50.757823] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.215 [2024-07-15 18:12:50.757829] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.215 [2024-07-15 18:12:50.757846] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.215 qpair failed and we were unable to recover it. 00:26:57.215 [2024-07-15 18:12:50.767707] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.215 [2024-07-15 18:12:50.767773] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.215 [2024-07-15 18:12:50.767787] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.215 [2024-07-15 18:12:50.767794] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.215 [2024-07-15 18:12:50.767800] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.215 [2024-07-15 18:12:50.767815] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.215 qpair failed and we were unable to recover it. 00:26:57.215 [2024-07-15 18:12:50.777763] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.215 [2024-07-15 18:12:50.777844] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.215 [2024-07-15 18:12:50.777859] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.215 [2024-07-15 18:12:50.777866] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.215 [2024-07-15 18:12:50.777872] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.215 [2024-07-15 18:12:50.777887] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.215 qpair failed and we were unable to recover it. 00:26:57.215 [2024-07-15 18:12:50.787790] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.215 [2024-07-15 18:12:50.787858] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.215 [2024-07-15 18:12:50.787875] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.215 [2024-07-15 18:12:50.787882] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.215 [2024-07-15 18:12:50.787888] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.215 [2024-07-15 18:12:50.787904] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.215 qpair failed and we were unable to recover it. 00:26:57.215 [2024-07-15 18:12:50.797806] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.215 [2024-07-15 18:12:50.797870] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.215 [2024-07-15 18:12:50.797885] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.215 [2024-07-15 18:12:50.797892] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.215 [2024-07-15 18:12:50.797898] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.215 [2024-07-15 18:12:50.797913] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.215 qpair failed and we were unable to recover it. 00:26:57.215 [2024-07-15 18:12:50.807855] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.215 [2024-07-15 18:12:50.807920] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.215 [2024-07-15 18:12:50.807940] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.215 [2024-07-15 18:12:50.807948] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.215 [2024-07-15 18:12:50.807954] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.215 [2024-07-15 18:12:50.807968] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.215 qpair failed and we were unable to recover it. 00:26:57.215 [2024-07-15 18:12:50.817859] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.215 [2024-07-15 18:12:50.817936] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.215 [2024-07-15 18:12:50.817950] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.215 [2024-07-15 18:12:50.817958] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.215 [2024-07-15 18:12:50.817964] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.215 [2024-07-15 18:12:50.817978] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.215 qpair failed and we were unable to recover it. 00:26:57.215 [2024-07-15 18:12:50.827880] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.215 [2024-07-15 18:12:50.827943] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.215 [2024-07-15 18:12:50.827958] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.215 [2024-07-15 18:12:50.827965] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.215 [2024-07-15 18:12:50.827971] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.215 [2024-07-15 18:12:50.827986] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.215 qpair failed and we were unable to recover it. 00:26:57.215 [2024-07-15 18:12:50.837933] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.216 [2024-07-15 18:12:50.838040] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.216 [2024-07-15 18:12:50.838055] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.216 [2024-07-15 18:12:50.838062] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.216 [2024-07-15 18:12:50.838069] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.216 [2024-07-15 18:12:50.838084] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.216 qpair failed and we were unable to recover it. 00:26:57.216 [2024-07-15 18:12:50.847935] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.216 [2024-07-15 18:12:50.847996] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.216 [2024-07-15 18:12:50.848011] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.216 [2024-07-15 18:12:50.848019] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.216 [2024-07-15 18:12:50.848028] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.216 [2024-07-15 18:12:50.848043] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.216 qpair failed and we were unable to recover it. 00:26:57.216 [2024-07-15 18:12:50.857990] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.216 [2024-07-15 18:12:50.858073] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.216 [2024-07-15 18:12:50.858088] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.216 [2024-07-15 18:12:50.858095] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.216 [2024-07-15 18:12:50.858101] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.216 [2024-07-15 18:12:50.858115] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.216 qpair failed and we were unable to recover it. 00:26:57.216 [2024-07-15 18:12:50.868062] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.216 [2024-07-15 18:12:50.868149] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.216 [2024-07-15 18:12:50.868164] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.216 [2024-07-15 18:12:50.868170] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.216 [2024-07-15 18:12:50.868176] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.216 [2024-07-15 18:12:50.868191] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.216 qpair failed and we were unable to recover it. 00:26:57.216 [2024-07-15 18:12:50.878017] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.216 [2024-07-15 18:12:50.878080] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.216 [2024-07-15 18:12:50.878095] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.216 [2024-07-15 18:12:50.878102] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.216 [2024-07-15 18:12:50.878108] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.216 [2024-07-15 18:12:50.878122] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.216 qpair failed and we were unable to recover it. 00:26:57.216 [2024-07-15 18:12:50.888035] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.216 [2024-07-15 18:12:50.888120] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.216 [2024-07-15 18:12:50.888135] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.216 [2024-07-15 18:12:50.888142] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.216 [2024-07-15 18:12:50.888149] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.216 [2024-07-15 18:12:50.888163] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.216 qpair failed and we were unable to recover it. 00:26:57.216 [2024-07-15 18:12:50.898104] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.216 [2024-07-15 18:12:50.898223] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.216 [2024-07-15 18:12:50.898242] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.216 [2024-07-15 18:12:50.898248] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.216 [2024-07-15 18:12:50.898255] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.216 [2024-07-15 18:12:50.898271] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.216 qpair failed and we were unable to recover it. 00:26:57.216 [2024-07-15 18:12:50.908122] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.216 [2024-07-15 18:12:50.908191] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.216 [2024-07-15 18:12:50.908206] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.216 [2024-07-15 18:12:50.908214] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.216 [2024-07-15 18:12:50.908221] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.216 [2024-07-15 18:12:50.908239] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.216 qpair failed and we were unable to recover it. 00:26:57.216 [2024-07-15 18:12:50.918122] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.216 [2024-07-15 18:12:50.918190] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.216 [2024-07-15 18:12:50.918205] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.216 [2024-07-15 18:12:50.918213] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.216 [2024-07-15 18:12:50.918219] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.216 [2024-07-15 18:12:50.918238] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.216 qpair failed and we were unable to recover it. 00:26:57.216 [2024-07-15 18:12:50.928156] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.216 [2024-07-15 18:12:50.928219] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.216 [2024-07-15 18:12:50.928236] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.216 [2024-07-15 18:12:50.928243] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.216 [2024-07-15 18:12:50.928249] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.216 [2024-07-15 18:12:50.928264] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.216 qpair failed and we were unable to recover it. 00:26:57.216 [2024-07-15 18:12:50.938181] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.216 [2024-07-15 18:12:50.938278] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.216 [2024-07-15 18:12:50.938293] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.216 [2024-07-15 18:12:50.938300] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.216 [2024-07-15 18:12:50.938309] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.216 [2024-07-15 18:12:50.938325] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.216 qpair failed and we were unable to recover it. 00:26:57.479 [2024-07-15 18:12:50.948206] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.479 [2024-07-15 18:12:50.948276] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.479 [2024-07-15 18:12:50.948292] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.479 [2024-07-15 18:12:50.948299] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.479 [2024-07-15 18:12:50.948306] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.479 [2024-07-15 18:12:50.948321] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.479 qpair failed and we were unable to recover it. 00:26:57.479 [2024-07-15 18:12:50.958220] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.479 [2024-07-15 18:12:50.958291] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.479 [2024-07-15 18:12:50.958308] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.479 [2024-07-15 18:12:50.958315] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.479 [2024-07-15 18:12:50.958322] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.479 [2024-07-15 18:12:50.958337] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.479 qpair failed and we were unable to recover it. 00:26:57.479 [2024-07-15 18:12:50.968268] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.479 [2024-07-15 18:12:50.968328] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.479 [2024-07-15 18:12:50.968342] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.479 [2024-07-15 18:12:50.968350] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.479 [2024-07-15 18:12:50.968356] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.479 [2024-07-15 18:12:50.968371] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.479 qpair failed and we were unable to recover it. 00:26:57.479 [2024-07-15 18:12:50.978293] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.479 [2024-07-15 18:12:50.978357] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.479 [2024-07-15 18:12:50.978371] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.479 [2024-07-15 18:12:50.978379] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.479 [2024-07-15 18:12:50.978385] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.479 [2024-07-15 18:12:50.978400] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.479 qpair failed and we were unable to recover it. 00:26:57.479 [2024-07-15 18:12:50.988325] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.479 [2024-07-15 18:12:50.988394] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.479 [2024-07-15 18:12:50.988410] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.479 [2024-07-15 18:12:50.988418] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.479 [2024-07-15 18:12:50.988424] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.479 [2024-07-15 18:12:50.988439] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.479 qpair failed and we were unable to recover it. 00:26:57.479 [2024-07-15 18:12:50.998327] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.479 [2024-07-15 18:12:50.998441] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.479 [2024-07-15 18:12:50.998457] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.479 [2024-07-15 18:12:50.998464] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.479 [2024-07-15 18:12:50.998471] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.479 [2024-07-15 18:12:50.998487] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.479 qpair failed and we were unable to recover it. 00:26:57.479 [2024-07-15 18:12:51.008363] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.479 [2024-07-15 18:12:51.008437] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.479 [2024-07-15 18:12:51.008452] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.479 [2024-07-15 18:12:51.008459] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.479 [2024-07-15 18:12:51.008465] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.479 [2024-07-15 18:12:51.008479] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.479 qpair failed and we were unable to recover it. 00:26:57.479 [2024-07-15 18:12:51.018361] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.479 [2024-07-15 18:12:51.018429] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.479 [2024-07-15 18:12:51.018444] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.479 [2024-07-15 18:12:51.018451] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.479 [2024-07-15 18:12:51.018457] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.479 [2024-07-15 18:12:51.018471] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.479 qpair failed and we were unable to recover it. 00:26:57.479 [2024-07-15 18:12:51.028427] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.479 [2024-07-15 18:12:51.028528] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.479 [2024-07-15 18:12:51.028543] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.479 [2024-07-15 18:12:51.028554] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.479 [2024-07-15 18:12:51.028560] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.479 [2024-07-15 18:12:51.028575] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.479 qpair failed and we were unable to recover it. 00:26:57.479 [2024-07-15 18:12:51.038491] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.479 [2024-07-15 18:12:51.038599] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.479 [2024-07-15 18:12:51.038614] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.480 [2024-07-15 18:12:51.038621] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.480 [2024-07-15 18:12:51.038628] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.480 [2024-07-15 18:12:51.038643] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.480 qpair failed and we were unable to recover it. 00:26:57.480 [2024-07-15 18:12:51.048501] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.480 [2024-07-15 18:12:51.048562] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.480 [2024-07-15 18:12:51.048576] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.480 [2024-07-15 18:12:51.048584] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.480 [2024-07-15 18:12:51.048590] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.480 [2024-07-15 18:12:51.048605] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.480 qpair failed and we were unable to recover it. 00:26:57.480 [2024-07-15 18:12:51.058536] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.480 [2024-07-15 18:12:51.058597] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.480 [2024-07-15 18:12:51.058612] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.480 [2024-07-15 18:12:51.058620] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.480 [2024-07-15 18:12:51.058626] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.480 [2024-07-15 18:12:51.058641] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.480 qpair failed and we were unable to recover it. 00:26:57.480 [2024-07-15 18:12:51.068613] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.480 [2024-07-15 18:12:51.068679] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.480 [2024-07-15 18:12:51.068694] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.480 [2024-07-15 18:12:51.068701] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.480 [2024-07-15 18:12:51.068707] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.480 [2024-07-15 18:12:51.068721] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.480 qpair failed and we were unable to recover it. 00:26:57.480 [2024-07-15 18:12:51.078580] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.480 [2024-07-15 18:12:51.078697] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.480 [2024-07-15 18:12:51.078714] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.480 [2024-07-15 18:12:51.078721] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.480 [2024-07-15 18:12:51.078728] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.480 [2024-07-15 18:12:51.078743] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.480 qpair failed and we were unable to recover it. 00:26:57.480 [2024-07-15 18:12:51.088653] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.480 [2024-07-15 18:12:51.088714] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.480 [2024-07-15 18:12:51.088729] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.480 [2024-07-15 18:12:51.088736] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.480 [2024-07-15 18:12:51.088743] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.480 [2024-07-15 18:12:51.088758] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.480 qpair failed and we were unable to recover it. 00:26:57.480 [2024-07-15 18:12:51.098667] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.480 [2024-07-15 18:12:51.098747] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.480 [2024-07-15 18:12:51.098763] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.480 [2024-07-15 18:12:51.098770] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.480 [2024-07-15 18:12:51.098776] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.480 [2024-07-15 18:12:51.098791] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.480 qpair failed and we were unable to recover it. 00:26:57.480 [2024-07-15 18:12:51.108697] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.480 [2024-07-15 18:12:51.108761] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.480 [2024-07-15 18:12:51.108776] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.480 [2024-07-15 18:12:51.108783] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.480 [2024-07-15 18:12:51.108790] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.480 [2024-07-15 18:12:51.108804] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.480 qpair failed and we were unable to recover it. 00:26:57.480 [2024-07-15 18:12:51.118764] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.480 [2024-07-15 18:12:51.118875] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.480 [2024-07-15 18:12:51.118895] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.480 [2024-07-15 18:12:51.118902] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.480 [2024-07-15 18:12:51.118908] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.480 [2024-07-15 18:12:51.118923] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.480 qpair failed and we were unable to recover it. 00:26:57.480 [2024-07-15 18:12:51.128722] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.480 [2024-07-15 18:12:51.128787] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.480 [2024-07-15 18:12:51.128801] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.480 [2024-07-15 18:12:51.128809] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.480 [2024-07-15 18:12:51.128815] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.480 [2024-07-15 18:12:51.128829] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.480 qpair failed and we were unable to recover it. 00:26:57.480 [2024-07-15 18:12:51.138801] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.480 [2024-07-15 18:12:51.138866] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.480 [2024-07-15 18:12:51.138880] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.480 [2024-07-15 18:12:51.138888] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.480 [2024-07-15 18:12:51.138894] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.480 [2024-07-15 18:12:51.138908] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.480 qpair failed and we were unable to recover it. 00:26:57.480 [2024-07-15 18:12:51.148858] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.480 [2024-07-15 18:12:51.148936] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.480 [2024-07-15 18:12:51.148951] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.480 [2024-07-15 18:12:51.148958] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.480 [2024-07-15 18:12:51.148964] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.480 [2024-07-15 18:12:51.148979] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.480 qpair failed and we were unable to recover it. 00:26:57.480 [2024-07-15 18:12:51.158837] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.480 [2024-07-15 18:12:51.158906] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.480 [2024-07-15 18:12:51.158921] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.480 [2024-07-15 18:12:51.158929] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.480 [2024-07-15 18:12:51.158935] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.480 [2024-07-15 18:12:51.158953] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.480 qpair failed and we were unable to recover it. 00:26:57.480 [2024-07-15 18:12:51.168865] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.480 [2024-07-15 18:12:51.168926] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.480 [2024-07-15 18:12:51.168941] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.480 [2024-07-15 18:12:51.168948] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.480 [2024-07-15 18:12:51.168954] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.480 [2024-07-15 18:12:51.168969] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.480 qpair failed and we were unable to recover it. 00:26:57.480 [2024-07-15 18:12:51.178849] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.480 [2024-07-15 18:12:51.178938] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.480 [2024-07-15 18:12:51.178952] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.480 [2024-07-15 18:12:51.178959] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.480 [2024-07-15 18:12:51.178965] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.480 [2024-07-15 18:12:51.178979] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.480 qpair failed and we were unable to recover it. 00:26:57.481 [2024-07-15 18:12:51.188896] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.481 [2024-07-15 18:12:51.188964] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.481 [2024-07-15 18:12:51.188980] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.481 [2024-07-15 18:12:51.188987] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.481 [2024-07-15 18:12:51.188993] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.481 [2024-07-15 18:12:51.189007] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.481 qpair failed and we were unable to recover it. 00:26:57.481 [2024-07-15 18:12:51.198923] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.481 [2024-07-15 18:12:51.198984] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.481 [2024-07-15 18:12:51.198999] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.481 [2024-07-15 18:12:51.199006] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.481 [2024-07-15 18:12:51.199013] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.481 [2024-07-15 18:12:51.199027] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.481 qpair failed and we were unable to recover it. 00:26:57.742 [2024-07-15 18:12:51.208941] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.742 [2024-07-15 18:12:51.209008] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.742 [2024-07-15 18:12:51.209026] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.742 [2024-07-15 18:12:51.209033] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.742 [2024-07-15 18:12:51.209039] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.742 [2024-07-15 18:12:51.209054] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.742 qpair failed and we were unable to recover it. 00:26:57.742 [2024-07-15 18:12:51.219034] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.742 [2024-07-15 18:12:51.219102] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.742 [2024-07-15 18:12:51.219117] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.742 [2024-07-15 18:12:51.219124] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.742 [2024-07-15 18:12:51.219130] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.742 [2024-07-15 18:12:51.219144] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.742 qpair failed and we were unable to recover it. 00:26:57.742 [2024-07-15 18:12:51.229031] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.742 [2024-07-15 18:12:51.229103] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.742 [2024-07-15 18:12:51.229118] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.742 [2024-07-15 18:12:51.229125] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.742 [2024-07-15 18:12:51.229131] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.742 [2024-07-15 18:12:51.229146] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.742 qpair failed and we were unable to recover it. 00:26:57.742 [2024-07-15 18:12:51.239084] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.742 [2024-07-15 18:12:51.239149] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.742 [2024-07-15 18:12:51.239164] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.742 [2024-07-15 18:12:51.239171] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.742 [2024-07-15 18:12:51.239177] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.742 [2024-07-15 18:12:51.239191] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.742 qpair failed and we were unable to recover it. 00:26:57.742 [2024-07-15 18:12:51.249064] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.742 [2024-07-15 18:12:51.249128] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.742 [2024-07-15 18:12:51.249143] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.742 [2024-07-15 18:12:51.249150] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.742 [2024-07-15 18:12:51.249156] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.742 [2024-07-15 18:12:51.249174] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.742 qpair failed and we were unable to recover it. 00:26:57.742 [2024-07-15 18:12:51.259118] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.742 [2024-07-15 18:12:51.259197] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.742 [2024-07-15 18:12:51.259212] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.742 [2024-07-15 18:12:51.259219] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.742 [2024-07-15 18:12:51.259230] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.742 [2024-07-15 18:12:51.259246] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.742 qpair failed and we were unable to recover it. 00:26:57.742 [2024-07-15 18:12:51.269169] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.742 [2024-07-15 18:12:51.269237] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.742 [2024-07-15 18:12:51.269253] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.742 [2024-07-15 18:12:51.269260] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.742 [2024-07-15 18:12:51.269267] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.742 [2024-07-15 18:12:51.269282] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.742 qpair failed and we were unable to recover it. 00:26:57.742 [2024-07-15 18:12:51.279194] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.742 [2024-07-15 18:12:51.279253] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.742 [2024-07-15 18:12:51.279268] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.742 [2024-07-15 18:12:51.279274] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.742 [2024-07-15 18:12:51.279281] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.742 [2024-07-15 18:12:51.279295] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.742 qpair failed and we were unable to recover it. 00:26:57.742 [2024-07-15 18:12:51.289192] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.742 [2024-07-15 18:12:51.289255] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.742 [2024-07-15 18:12:51.289270] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.742 [2024-07-15 18:12:51.289277] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.742 [2024-07-15 18:12:51.289283] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.742 [2024-07-15 18:12:51.289298] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.742 qpair failed and we were unable to recover it. 00:26:57.742 [2024-07-15 18:12:51.299209] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.742 [2024-07-15 18:12:51.299282] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.742 [2024-07-15 18:12:51.299297] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.742 [2024-07-15 18:12:51.299304] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.742 [2024-07-15 18:12:51.299310] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.743 [2024-07-15 18:12:51.299324] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.743 qpair failed and we were unable to recover it. 00:26:57.743 [2024-07-15 18:12:51.309217] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.743 [2024-07-15 18:12:51.309336] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.743 [2024-07-15 18:12:51.309350] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.743 [2024-07-15 18:12:51.309358] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.743 [2024-07-15 18:12:51.309364] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.743 [2024-07-15 18:12:51.309379] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.743 qpair failed and we were unable to recover it. 00:26:57.743 [2024-07-15 18:12:51.319257] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.743 [2024-07-15 18:12:51.319357] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.743 [2024-07-15 18:12:51.319372] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.743 [2024-07-15 18:12:51.319379] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.743 [2024-07-15 18:12:51.319385] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.743 [2024-07-15 18:12:51.319401] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.743 qpair failed and we were unable to recover it. 00:26:57.743 [2024-07-15 18:12:51.329273] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.743 [2024-07-15 18:12:51.329357] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.743 [2024-07-15 18:12:51.329372] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.743 [2024-07-15 18:12:51.329379] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.743 [2024-07-15 18:12:51.329386] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.743 [2024-07-15 18:12:51.329400] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.743 qpair failed and we were unable to recover it. 00:26:57.743 [2024-07-15 18:12:51.339349] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.743 [2024-07-15 18:12:51.339418] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.743 [2024-07-15 18:12:51.339433] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.743 [2024-07-15 18:12:51.339440] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.743 [2024-07-15 18:12:51.339449] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.743 [2024-07-15 18:12:51.339465] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.743 qpair failed and we were unable to recover it. 00:26:57.743 [2024-07-15 18:12:51.349341] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.743 [2024-07-15 18:12:51.349402] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.743 [2024-07-15 18:12:51.349416] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.743 [2024-07-15 18:12:51.349423] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.743 [2024-07-15 18:12:51.349429] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.743 [2024-07-15 18:12:51.349443] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.743 qpair failed and we were unable to recover it. 00:26:57.743 [2024-07-15 18:12:51.359387] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.743 [2024-07-15 18:12:51.359456] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.743 [2024-07-15 18:12:51.359470] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.743 [2024-07-15 18:12:51.359477] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.743 [2024-07-15 18:12:51.359483] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.743 [2024-07-15 18:12:51.359497] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.743 qpair failed and we were unable to recover it. 00:26:57.743 [2024-07-15 18:12:51.369436] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.743 [2024-07-15 18:12:51.369540] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.743 [2024-07-15 18:12:51.369556] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.743 [2024-07-15 18:12:51.369563] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.743 [2024-07-15 18:12:51.369570] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.743 [2024-07-15 18:12:51.369585] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.743 qpair failed and we were unable to recover it. 00:26:57.743 [2024-07-15 18:12:51.379440] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.743 [2024-07-15 18:12:51.379505] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.743 [2024-07-15 18:12:51.379520] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.743 [2024-07-15 18:12:51.379527] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.743 [2024-07-15 18:12:51.379534] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.743 [2024-07-15 18:12:51.379548] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.743 qpair failed and we were unable to recover it. 00:26:57.743 [2024-07-15 18:12:51.389419] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.743 [2024-07-15 18:12:51.389488] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.743 [2024-07-15 18:12:51.389503] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.743 [2024-07-15 18:12:51.389510] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.743 [2024-07-15 18:12:51.389516] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.743 [2024-07-15 18:12:51.389531] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.743 qpair failed and we were unable to recover it. 00:26:57.743 [2024-07-15 18:12:51.399490] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.743 [2024-07-15 18:12:51.399551] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.743 [2024-07-15 18:12:51.399566] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.743 [2024-07-15 18:12:51.399573] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.743 [2024-07-15 18:12:51.399579] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.743 [2024-07-15 18:12:51.399593] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.743 qpair failed and we were unable to recover it. 00:26:57.743 [2024-07-15 18:12:51.409603] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.743 [2024-07-15 18:12:51.409666] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.743 [2024-07-15 18:12:51.409681] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.743 [2024-07-15 18:12:51.409689] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.743 [2024-07-15 18:12:51.409695] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.743 [2024-07-15 18:12:51.409709] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.743 qpair failed and we were unable to recover it. 00:26:57.743 [2024-07-15 18:12:51.419611] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.743 [2024-07-15 18:12:51.419685] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.743 [2024-07-15 18:12:51.419700] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.743 [2024-07-15 18:12:51.419707] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.743 [2024-07-15 18:12:51.419713] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.743 [2024-07-15 18:12:51.419727] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.743 qpair failed and we were unable to recover it. 00:26:57.743 [2024-07-15 18:12:51.429594] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.743 [2024-07-15 18:12:51.429662] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.743 [2024-07-15 18:12:51.429676] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.743 [2024-07-15 18:12:51.429686] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.743 [2024-07-15 18:12:51.429692] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.743 [2024-07-15 18:12:51.429707] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.743 qpair failed and we were unable to recover it. 00:26:57.743 [2024-07-15 18:12:51.439613] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.743 [2024-07-15 18:12:51.439690] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.743 [2024-07-15 18:12:51.439706] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.743 [2024-07-15 18:12:51.439713] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.743 [2024-07-15 18:12:51.439720] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.743 [2024-07-15 18:12:51.439735] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.743 qpair failed and we were unable to recover it. 00:26:57.743 [2024-07-15 18:12:51.449637] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.744 [2024-07-15 18:12:51.449698] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.744 [2024-07-15 18:12:51.449712] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.744 [2024-07-15 18:12:51.449720] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.744 [2024-07-15 18:12:51.449726] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.744 [2024-07-15 18:12:51.449741] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.744 qpair failed and we were unable to recover it. 00:26:57.744 [2024-07-15 18:12:51.459685] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:57.744 [2024-07-15 18:12:51.459752] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:57.744 [2024-07-15 18:12:51.459767] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:57.744 [2024-07-15 18:12:51.459774] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:57.744 [2024-07-15 18:12:51.459780] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:57.744 [2024-07-15 18:12:51.459795] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:57.744 qpair failed and we were unable to recover it. 00:26:58.005 [2024-07-15 18:12:51.469699] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.005 [2024-07-15 18:12:51.469763] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.005 [2024-07-15 18:12:51.469777] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.005 [2024-07-15 18:12:51.469784] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.005 [2024-07-15 18:12:51.469790] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.005 [2024-07-15 18:12:51.469805] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.005 qpair failed and we were unable to recover it. 00:26:58.005 [2024-07-15 18:12:51.479730] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.005 [2024-07-15 18:12:51.479790] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.005 [2024-07-15 18:12:51.479805] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.005 [2024-07-15 18:12:51.479813] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.005 [2024-07-15 18:12:51.479819] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.005 [2024-07-15 18:12:51.479833] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.005 qpair failed and we were unable to recover it. 00:26:58.005 [2024-07-15 18:12:51.489797] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.005 [2024-07-15 18:12:51.489859] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.005 [2024-07-15 18:12:51.489874] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.005 [2024-07-15 18:12:51.489881] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.005 [2024-07-15 18:12:51.489887] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.005 [2024-07-15 18:12:51.489902] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.005 qpair failed and we were unable to recover it. 00:26:58.005 [2024-07-15 18:12:51.499813] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.005 [2024-07-15 18:12:51.499892] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.005 [2024-07-15 18:12:51.499907] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.005 [2024-07-15 18:12:51.499915] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.005 [2024-07-15 18:12:51.499921] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.005 [2024-07-15 18:12:51.499936] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.005 qpair failed and we were unable to recover it. 00:26:58.005 [2024-07-15 18:12:51.509767] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.005 [2024-07-15 18:12:51.509833] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.005 [2024-07-15 18:12:51.509847] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.006 [2024-07-15 18:12:51.509854] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.006 [2024-07-15 18:12:51.509861] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.006 [2024-07-15 18:12:51.509875] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.006 qpair failed and we were unable to recover it. 00:26:58.006 [2024-07-15 18:12:51.519792] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.006 [2024-07-15 18:12:51.519852] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.006 [2024-07-15 18:12:51.519866] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.006 [2024-07-15 18:12:51.519876] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.006 [2024-07-15 18:12:51.519882] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.006 [2024-07-15 18:12:51.519896] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.006 qpair failed and we were unable to recover it. 00:26:58.006 [2024-07-15 18:12:51.529924] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.006 [2024-07-15 18:12:51.529987] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.006 [2024-07-15 18:12:51.530001] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.006 [2024-07-15 18:12:51.530009] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.006 [2024-07-15 18:12:51.530015] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.006 [2024-07-15 18:12:51.530030] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.006 qpair failed and we were unable to recover it. 00:26:58.006 [2024-07-15 18:12:51.539938] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.006 [2024-07-15 18:12:51.540005] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.006 [2024-07-15 18:12:51.540020] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.006 [2024-07-15 18:12:51.540027] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.006 [2024-07-15 18:12:51.540033] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.006 [2024-07-15 18:12:51.540047] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.006 qpair failed and we were unable to recover it. 00:26:58.006 [2024-07-15 18:12:51.549947] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.006 [2024-07-15 18:12:51.550029] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.006 [2024-07-15 18:12:51.550044] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.006 [2024-07-15 18:12:51.550051] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.006 [2024-07-15 18:12:51.550057] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.006 [2024-07-15 18:12:51.550071] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.006 qpair failed and we were unable to recover it. 00:26:58.006 [2024-07-15 18:12:51.559965] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.006 [2024-07-15 18:12:51.560032] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.006 [2024-07-15 18:12:51.560046] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.006 [2024-07-15 18:12:51.560053] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.006 [2024-07-15 18:12:51.560060] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.006 [2024-07-15 18:12:51.560074] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.006 qpair failed and we were unable to recover it. 00:26:58.006 [2024-07-15 18:12:51.570001] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.006 [2024-07-15 18:12:51.570068] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.006 [2024-07-15 18:12:51.570085] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.006 [2024-07-15 18:12:51.570092] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.006 [2024-07-15 18:12:51.570098] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.006 [2024-07-15 18:12:51.570113] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.006 qpair failed and we were unable to recover it. 00:26:58.006 [2024-07-15 18:12:51.580014] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.006 [2024-07-15 18:12:51.580112] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.006 [2024-07-15 18:12:51.580127] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.006 [2024-07-15 18:12:51.580134] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.006 [2024-07-15 18:12:51.580140] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.006 [2024-07-15 18:12:51.580155] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.006 qpair failed and we were unable to recover it. 00:26:58.006 [2024-07-15 18:12:51.590050] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.006 [2024-07-15 18:12:51.590119] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.006 [2024-07-15 18:12:51.590136] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.006 [2024-07-15 18:12:51.590143] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.006 [2024-07-15 18:12:51.590149] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.006 [2024-07-15 18:12:51.590165] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.006 qpair failed and we were unable to recover it. 00:26:58.006 [2024-07-15 18:12:51.600047] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.006 [2024-07-15 18:12:51.600121] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.006 [2024-07-15 18:12:51.600137] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.006 [2024-07-15 18:12:51.600144] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.006 [2024-07-15 18:12:51.600150] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.006 [2024-07-15 18:12:51.600165] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.006 qpair failed and we were unable to recover it. 00:26:58.006 [2024-07-15 18:12:51.610142] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.006 [2024-07-15 18:12:51.610208] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.006 [2024-07-15 18:12:51.610232] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.006 [2024-07-15 18:12:51.610240] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.006 [2024-07-15 18:12:51.610246] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.006 [2024-07-15 18:12:51.610261] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.006 qpair failed and we were unable to recover it. 00:26:58.006 [2024-07-15 18:12:51.620127] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.006 [2024-07-15 18:12:51.620191] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.006 [2024-07-15 18:12:51.620207] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.006 [2024-07-15 18:12:51.620214] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.006 [2024-07-15 18:12:51.620220] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.006 [2024-07-15 18:12:51.620239] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.006 qpair failed and we were unable to recover it. 00:26:58.006 [2024-07-15 18:12:51.630201] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.006 [2024-07-15 18:12:51.630277] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.006 [2024-07-15 18:12:51.630292] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.006 [2024-07-15 18:12:51.630300] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.006 [2024-07-15 18:12:51.630306] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.006 [2024-07-15 18:12:51.630321] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.006 qpair failed and we were unable to recover it. 00:26:58.006 [2024-07-15 18:12:51.640121] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.006 [2024-07-15 18:12:51.640191] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.006 [2024-07-15 18:12:51.640205] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.006 [2024-07-15 18:12:51.640212] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.006 [2024-07-15 18:12:51.640218] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.006 [2024-07-15 18:12:51.640237] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.006 qpair failed and we were unable to recover it. 00:26:58.006 [2024-07-15 18:12:51.650254] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.006 [2024-07-15 18:12:51.650317] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.006 [2024-07-15 18:12:51.650333] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.006 [2024-07-15 18:12:51.650341] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.006 [2024-07-15 18:12:51.650347] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.006 [2024-07-15 18:12:51.650365] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.007 qpair failed and we were unable to recover it. 00:26:58.007 [2024-07-15 18:12:51.660253] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.007 [2024-07-15 18:12:51.660318] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.007 [2024-07-15 18:12:51.660333] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.007 [2024-07-15 18:12:51.660340] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.007 [2024-07-15 18:12:51.660346] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.007 [2024-07-15 18:12:51.660361] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.007 qpair failed and we were unable to recover it. 00:26:58.007 [2024-07-15 18:12:51.670284] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.007 [2024-07-15 18:12:51.670389] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.007 [2024-07-15 18:12:51.670404] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.007 [2024-07-15 18:12:51.670411] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.007 [2024-07-15 18:12:51.670418] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.007 [2024-07-15 18:12:51.670435] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.007 qpair failed and we were unable to recover it. 00:26:58.007 [2024-07-15 18:12:51.680374] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.007 [2024-07-15 18:12:51.680473] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.007 [2024-07-15 18:12:51.680488] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.007 [2024-07-15 18:12:51.680495] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.007 [2024-07-15 18:12:51.680501] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.007 [2024-07-15 18:12:51.680517] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.007 qpair failed and we were unable to recover it. 00:26:58.007 [2024-07-15 18:12:51.690341] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.007 [2024-07-15 18:12:51.690407] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.007 [2024-07-15 18:12:51.690422] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.007 [2024-07-15 18:12:51.690429] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.007 [2024-07-15 18:12:51.690436] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.007 [2024-07-15 18:12:51.690451] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.007 qpair failed and we were unable to recover it. 00:26:58.007 [2024-07-15 18:12:51.700389] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.007 [2024-07-15 18:12:51.700455] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.007 [2024-07-15 18:12:51.700473] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.007 [2024-07-15 18:12:51.700480] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.007 [2024-07-15 18:12:51.700487] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.007 [2024-07-15 18:12:51.700502] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.007 qpair failed and we were unable to recover it. 00:26:58.007 [2024-07-15 18:12:51.710421] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.007 [2024-07-15 18:12:51.710487] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.007 [2024-07-15 18:12:51.710502] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.007 [2024-07-15 18:12:51.710509] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.007 [2024-07-15 18:12:51.710516] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.007 [2024-07-15 18:12:51.710531] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.007 qpair failed and we were unable to recover it. 00:26:58.007 [2024-07-15 18:12:51.720421] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.007 [2024-07-15 18:12:51.720532] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.007 [2024-07-15 18:12:51.720548] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.007 [2024-07-15 18:12:51.720556] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.007 [2024-07-15 18:12:51.720563] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.007 [2024-07-15 18:12:51.720577] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.007 qpair failed and we were unable to recover it. 00:26:58.267 [2024-07-15 18:12:51.730397] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.267 [2024-07-15 18:12:51.730480] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.267 [2024-07-15 18:12:51.730494] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.267 [2024-07-15 18:12:51.730501] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.267 [2024-07-15 18:12:51.730507] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.267 [2024-07-15 18:12:51.730522] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.267 qpair failed and we were unable to recover it. 00:26:58.267 [2024-07-15 18:12:51.740466] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.267 [2024-07-15 18:12:51.740529] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.267 [2024-07-15 18:12:51.740544] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.267 [2024-07-15 18:12:51.740551] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.267 [2024-07-15 18:12:51.740561] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.267 [2024-07-15 18:12:51.740575] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.267 qpair failed and we were unable to recover it. 00:26:58.267 [2024-07-15 18:12:51.750483] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.267 [2024-07-15 18:12:51.750551] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.267 [2024-07-15 18:12:51.750566] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.267 [2024-07-15 18:12:51.750573] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.267 [2024-07-15 18:12:51.750579] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.267 [2024-07-15 18:12:51.750594] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.267 qpair failed and we were unable to recover it. 00:26:58.267 [2024-07-15 18:12:51.760539] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.267 [2024-07-15 18:12:51.760601] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.267 [2024-07-15 18:12:51.760616] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.267 [2024-07-15 18:12:51.760623] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.267 [2024-07-15 18:12:51.760630] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.267 [2024-07-15 18:12:51.760645] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.267 qpair failed and we were unable to recover it. 00:26:58.267 [2024-07-15 18:12:51.770565] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.267 [2024-07-15 18:12:51.770643] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.267 [2024-07-15 18:12:51.770658] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.267 [2024-07-15 18:12:51.770665] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.267 [2024-07-15 18:12:51.770671] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.267 [2024-07-15 18:12:51.770686] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.267 qpair failed and we were unable to recover it. 00:26:58.267 [2024-07-15 18:12:51.780599] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.267 [2024-07-15 18:12:51.780676] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.267 [2024-07-15 18:12:51.780692] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.267 [2024-07-15 18:12:51.780700] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.267 [2024-07-15 18:12:51.780707] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.267 [2024-07-15 18:12:51.780722] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.267 qpair failed and we were unable to recover it. 00:26:58.267 [2024-07-15 18:12:51.790682] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.267 [2024-07-15 18:12:51.790790] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.267 [2024-07-15 18:12:51.790806] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.267 [2024-07-15 18:12:51.790813] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.267 [2024-07-15 18:12:51.790819] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.267 [2024-07-15 18:12:51.790833] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.267 qpair failed and we were unable to recover it. 00:26:58.267 [2024-07-15 18:12:51.800654] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.267 [2024-07-15 18:12:51.800715] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.267 [2024-07-15 18:12:51.800730] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.267 [2024-07-15 18:12:51.800737] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.267 [2024-07-15 18:12:51.800743] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.267 [2024-07-15 18:12:51.800759] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.267 qpair failed and we were unable to recover it. 00:26:58.267 [2024-07-15 18:12:51.810690] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.267 [2024-07-15 18:12:51.810751] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.267 [2024-07-15 18:12:51.810766] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.267 [2024-07-15 18:12:51.810774] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.267 [2024-07-15 18:12:51.810780] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.267 [2024-07-15 18:12:51.810795] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.267 qpair failed and we were unable to recover it. 00:26:58.267 [2024-07-15 18:12:51.820721] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.267 [2024-07-15 18:12:51.820794] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.267 [2024-07-15 18:12:51.820809] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.268 [2024-07-15 18:12:51.820816] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.268 [2024-07-15 18:12:51.820822] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.268 [2024-07-15 18:12:51.820837] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.268 qpair failed and we were unable to recover it. 00:26:58.268 [2024-07-15 18:12:51.830731] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.268 [2024-07-15 18:12:51.830795] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.268 [2024-07-15 18:12:51.830811] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.268 [2024-07-15 18:12:51.830821] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.268 [2024-07-15 18:12:51.830828] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.268 [2024-07-15 18:12:51.830843] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.268 qpair failed and we were unable to recover it. 00:26:58.268 [2024-07-15 18:12:51.840730] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.268 [2024-07-15 18:12:51.840795] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.268 [2024-07-15 18:12:51.840809] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.268 [2024-07-15 18:12:51.840816] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.268 [2024-07-15 18:12:51.840822] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.268 [2024-07-15 18:12:51.840837] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.268 qpair failed and we were unable to recover it. 00:26:58.268 [2024-07-15 18:12:51.850825] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.268 [2024-07-15 18:12:51.850904] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.268 [2024-07-15 18:12:51.850919] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.268 [2024-07-15 18:12:51.850926] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.268 [2024-07-15 18:12:51.850933] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.268 [2024-07-15 18:12:51.850947] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.268 qpair failed and we were unable to recover it. 00:26:58.268 [2024-07-15 18:12:51.860828] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.268 [2024-07-15 18:12:51.860896] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.268 [2024-07-15 18:12:51.860911] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.268 [2024-07-15 18:12:51.860918] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.268 [2024-07-15 18:12:51.860924] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.268 [2024-07-15 18:12:51.860939] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.268 qpair failed and we were unable to recover it. 00:26:58.268 [2024-07-15 18:12:51.870826] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.268 [2024-07-15 18:12:51.870898] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.268 [2024-07-15 18:12:51.870915] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.268 [2024-07-15 18:12:51.870923] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.268 [2024-07-15 18:12:51.870929] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.268 [2024-07-15 18:12:51.870945] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.268 qpair failed and we were unable to recover it. 00:26:58.268 [2024-07-15 18:12:51.880904] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.268 [2024-07-15 18:12:51.880971] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.268 [2024-07-15 18:12:51.880986] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.268 [2024-07-15 18:12:51.880993] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.268 [2024-07-15 18:12:51.880999] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.268 [2024-07-15 18:12:51.881013] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.268 qpair failed and we were unable to recover it. 00:26:58.268 [2024-07-15 18:12:51.890972] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.268 [2024-07-15 18:12:51.891034] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.268 [2024-07-15 18:12:51.891049] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.268 [2024-07-15 18:12:51.891056] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.268 [2024-07-15 18:12:51.891062] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.268 [2024-07-15 18:12:51.891077] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.268 qpair failed and we were unable to recover it. 00:26:58.268 [2024-07-15 18:12:51.900875] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.268 [2024-07-15 18:12:51.900938] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.268 [2024-07-15 18:12:51.900953] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.268 [2024-07-15 18:12:51.900961] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.268 [2024-07-15 18:12:51.900967] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.268 [2024-07-15 18:12:51.900982] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.268 qpair failed and we were unable to recover it. 00:26:58.268 [2024-07-15 18:12:51.911026] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.268 [2024-07-15 18:12:51.911090] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.268 [2024-07-15 18:12:51.911105] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.268 [2024-07-15 18:12:51.911112] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.268 [2024-07-15 18:12:51.911119] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.268 [2024-07-15 18:12:51.911134] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.268 qpair failed and we were unable to recover it. 00:26:58.268 [2024-07-15 18:12:51.920992] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.268 [2024-07-15 18:12:51.921050] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.268 [2024-07-15 18:12:51.921065] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.268 [2024-07-15 18:12:51.921076] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.268 [2024-07-15 18:12:51.921082] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.268 [2024-07-15 18:12:51.921097] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.268 qpair failed and we were unable to recover it. 00:26:58.268 [2024-07-15 18:12:51.931053] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.268 [2024-07-15 18:12:51.931162] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.268 [2024-07-15 18:12:51.931185] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.268 [2024-07-15 18:12:51.931192] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.268 [2024-07-15 18:12:51.931198] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.268 [2024-07-15 18:12:51.931214] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.268 qpair failed and we were unable to recover it. 00:26:58.268 [2024-07-15 18:12:51.941036] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.268 [2024-07-15 18:12:51.941101] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.268 [2024-07-15 18:12:51.941116] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.268 [2024-07-15 18:12:51.941123] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.268 [2024-07-15 18:12:51.941129] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.268 [2024-07-15 18:12:51.941144] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.268 qpair failed and we were unable to recover it. 00:26:58.268 [2024-07-15 18:12:51.951058] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.268 [2024-07-15 18:12:51.951165] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.268 [2024-07-15 18:12:51.951180] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.268 [2024-07-15 18:12:51.951188] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.268 [2024-07-15 18:12:51.951195] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.268 [2024-07-15 18:12:51.951211] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.268 qpair failed and we were unable to recover it. 00:26:58.268 [2024-07-15 18:12:51.961114] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.268 [2024-07-15 18:12:51.961182] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.268 [2024-07-15 18:12:51.961198] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.268 [2024-07-15 18:12:51.961205] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.268 [2024-07-15 18:12:51.961212] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.269 [2024-07-15 18:12:51.961230] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.269 qpair failed and we were unable to recover it. 00:26:58.269 [2024-07-15 18:12:51.971177] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.269 [2024-07-15 18:12:51.971246] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.269 [2024-07-15 18:12:51.971262] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.269 [2024-07-15 18:12:51.971270] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.269 [2024-07-15 18:12:51.971276] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.269 [2024-07-15 18:12:51.971292] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.269 qpair failed and we were unable to recover it. 00:26:58.269 [2024-07-15 18:12:51.981174] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.269 [2024-07-15 18:12:51.981240] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.269 [2024-07-15 18:12:51.981254] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.269 [2024-07-15 18:12:51.981262] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.269 [2024-07-15 18:12:51.981269] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.269 [2024-07-15 18:12:51.981283] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.269 qpair failed and we were unable to recover it. 00:26:58.269 [2024-07-15 18:12:51.991236] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.269 [2024-07-15 18:12:51.991302] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.269 [2024-07-15 18:12:51.991317] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.269 [2024-07-15 18:12:51.991325] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.269 [2024-07-15 18:12:51.991331] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.269 [2024-07-15 18:12:51.991347] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.269 qpair failed and we were unable to recover it. 00:26:58.528 [2024-07-15 18:12:52.001231] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.528 [2024-07-15 18:12:52.001346] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.528 [2024-07-15 18:12:52.001364] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.528 [2024-07-15 18:12:52.001373] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.528 [2024-07-15 18:12:52.001382] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.528 [2024-07-15 18:12:52.001398] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.528 qpair failed and we were unable to recover it. 00:26:58.528 [2024-07-15 18:12:52.011270] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.528 [2024-07-15 18:12:52.011340] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.528 [2024-07-15 18:12:52.011359] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.528 [2024-07-15 18:12:52.011367] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.528 [2024-07-15 18:12:52.011373] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.528 [2024-07-15 18:12:52.011388] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.528 qpair failed and we were unable to recover it. 00:26:58.528 [2024-07-15 18:12:52.021286] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.528 [2024-07-15 18:12:52.021353] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.528 [2024-07-15 18:12:52.021369] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.528 [2024-07-15 18:12:52.021376] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.528 [2024-07-15 18:12:52.021382] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.528 [2024-07-15 18:12:52.021397] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.528 qpair failed and we were unable to recover it. 00:26:58.528 [2024-07-15 18:12:52.031296] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.528 [2024-07-15 18:12:52.031364] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.528 [2024-07-15 18:12:52.031378] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.528 [2024-07-15 18:12:52.031385] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.528 [2024-07-15 18:12:52.031391] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.528 [2024-07-15 18:12:52.031406] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.528 qpair failed and we were unable to recover it. 00:26:58.528 [2024-07-15 18:12:52.041332] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.528 [2024-07-15 18:12:52.041398] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.528 [2024-07-15 18:12:52.041413] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.528 [2024-07-15 18:12:52.041420] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.528 [2024-07-15 18:12:52.041426] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.528 [2024-07-15 18:12:52.041440] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.528 qpair failed and we were unable to recover it. 00:26:58.528 [2024-07-15 18:12:52.051379] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.528 [2024-07-15 18:12:52.051457] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.528 [2024-07-15 18:12:52.051472] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.528 [2024-07-15 18:12:52.051479] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.528 [2024-07-15 18:12:52.051485] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.528 [2024-07-15 18:12:52.051502] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.528 qpair failed and we were unable to recover it. 00:26:58.528 [2024-07-15 18:12:52.061408] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.528 [2024-07-15 18:12:52.061476] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.528 [2024-07-15 18:12:52.061491] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.528 [2024-07-15 18:12:52.061498] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.528 [2024-07-15 18:12:52.061504] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.528 [2024-07-15 18:12:52.061519] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.528 qpair failed and we were unable to recover it. 00:26:58.528 [2024-07-15 18:12:52.071445] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.528 [2024-07-15 18:12:52.071508] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.528 [2024-07-15 18:12:52.071522] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.528 [2024-07-15 18:12:52.071530] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.528 [2024-07-15 18:12:52.071537] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.528 [2024-07-15 18:12:52.071552] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.528 qpair failed and we were unable to recover it. 00:26:58.528 [2024-07-15 18:12:52.081486] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.528 [2024-07-15 18:12:52.081546] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.528 [2024-07-15 18:12:52.081562] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.528 [2024-07-15 18:12:52.081570] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.528 [2024-07-15 18:12:52.081577] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:26:58.528 [2024-07-15 18:12:52.081592] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:26:58.528 qpair failed and we were unable to recover it. 00:26:58.528 [2024-07-15 18:12:52.091519] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.528 [2024-07-15 18:12:52.091643] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.528 [2024-07-15 18:12:52.091673] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.528 [2024-07-15 18:12:52.091685] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.528 [2024-07-15 18:12:52.091695] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.528 [2024-07-15 18:12:52.091717] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.528 qpair failed and we were unable to recover it. 00:26:58.528 [2024-07-15 18:12:52.101561] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.528 [2024-07-15 18:12:52.101628] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.528 [2024-07-15 18:12:52.101658] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.528 [2024-07-15 18:12:52.101666] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.528 [2024-07-15 18:12:52.101672] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.528 [2024-07-15 18:12:52.101689] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.528 qpair failed and we were unable to recover it. 00:26:58.528 [2024-07-15 18:12:52.111613] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.528 [2024-07-15 18:12:52.111726] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.528 [2024-07-15 18:12:52.111743] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.528 [2024-07-15 18:12:52.111750] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.528 [2024-07-15 18:12:52.111756] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.528 [2024-07-15 18:12:52.111771] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.528 qpair failed and we were unable to recover it. 00:26:58.528 [2024-07-15 18:12:52.121586] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.528 [2024-07-15 18:12:52.121648] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.528 [2024-07-15 18:12:52.121664] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.528 [2024-07-15 18:12:52.121672] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.528 [2024-07-15 18:12:52.121678] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.529 [2024-07-15 18:12:52.121692] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.529 qpair failed and we were unable to recover it. 00:26:58.529 [2024-07-15 18:12:52.131584] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.529 [2024-07-15 18:12:52.131649] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.529 [2024-07-15 18:12:52.131665] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.529 [2024-07-15 18:12:52.131672] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.529 [2024-07-15 18:12:52.131678] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.529 [2024-07-15 18:12:52.131692] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.529 qpair failed and we were unable to recover it. 00:26:58.529 [2024-07-15 18:12:52.141667] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.529 [2024-07-15 18:12:52.141731] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.529 [2024-07-15 18:12:52.141747] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.529 [2024-07-15 18:12:52.141754] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.529 [2024-07-15 18:12:52.141763] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.529 [2024-07-15 18:12:52.141778] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.529 qpair failed and we were unable to recover it. 00:26:58.529 [2024-07-15 18:12:52.151677] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.529 [2024-07-15 18:12:52.151775] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.529 [2024-07-15 18:12:52.151791] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.529 [2024-07-15 18:12:52.151798] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.529 [2024-07-15 18:12:52.151805] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.529 [2024-07-15 18:12:52.151820] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.529 qpair failed and we were unable to recover it. 00:26:58.529 [2024-07-15 18:12:52.161668] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.529 [2024-07-15 18:12:52.161733] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.529 [2024-07-15 18:12:52.161750] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.529 [2024-07-15 18:12:52.161757] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.529 [2024-07-15 18:12:52.161764] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.529 [2024-07-15 18:12:52.161778] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.529 qpair failed and we were unable to recover it. 00:26:58.529 [2024-07-15 18:12:52.171733] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.529 [2024-07-15 18:12:52.171808] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.529 [2024-07-15 18:12:52.171824] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.529 [2024-07-15 18:12:52.171831] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.529 [2024-07-15 18:12:52.171838] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.529 [2024-07-15 18:12:52.171852] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.529 qpair failed and we were unable to recover it. 00:26:58.529 [2024-07-15 18:12:52.181756] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.529 [2024-07-15 18:12:52.181825] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.529 [2024-07-15 18:12:52.181841] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.529 [2024-07-15 18:12:52.181848] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.529 [2024-07-15 18:12:52.181854] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.529 [2024-07-15 18:12:52.181868] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.529 qpair failed and we were unable to recover it. 00:26:58.529 [2024-07-15 18:12:52.191777] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.529 [2024-07-15 18:12:52.191849] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.529 [2024-07-15 18:12:52.191865] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.529 [2024-07-15 18:12:52.191872] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.529 [2024-07-15 18:12:52.191879] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.529 [2024-07-15 18:12:52.191893] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.529 qpair failed and we were unable to recover it. 00:26:58.529 [2024-07-15 18:12:52.201799] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.529 [2024-07-15 18:12:52.201892] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.529 [2024-07-15 18:12:52.201907] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.529 [2024-07-15 18:12:52.201914] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.529 [2024-07-15 18:12:52.201920] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.529 [2024-07-15 18:12:52.201935] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.529 qpair failed and we were unable to recover it. 00:26:58.529 [2024-07-15 18:12:52.211806] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.529 [2024-07-15 18:12:52.211872] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.529 [2024-07-15 18:12:52.211889] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.529 [2024-07-15 18:12:52.211897] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.529 [2024-07-15 18:12:52.211903] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.529 [2024-07-15 18:12:52.211917] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.529 qpair failed and we were unable to recover it. 00:26:58.529 [2024-07-15 18:12:52.221907] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.529 [2024-07-15 18:12:52.221969] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.529 [2024-07-15 18:12:52.221985] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.529 [2024-07-15 18:12:52.221992] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.529 [2024-07-15 18:12:52.221998] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.529 [2024-07-15 18:12:52.222012] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.529 qpair failed and we were unable to recover it. 00:26:58.529 [2024-07-15 18:12:52.231916] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.529 [2024-07-15 18:12:52.232025] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.529 [2024-07-15 18:12:52.232043] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.529 [2024-07-15 18:12:52.232050] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.529 [2024-07-15 18:12:52.232060] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.529 [2024-07-15 18:12:52.232076] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.529 qpair failed and we were unable to recover it. 00:26:58.529 [2024-07-15 18:12:52.241928] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.529 [2024-07-15 18:12:52.241988] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.529 [2024-07-15 18:12:52.242004] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.529 [2024-07-15 18:12:52.242011] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.529 [2024-07-15 18:12:52.242018] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.529 [2024-07-15 18:12:52.242032] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.529 qpair failed and we were unable to recover it. 00:26:58.529 [2024-07-15 18:12:52.251946] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.529 [2024-07-15 18:12:52.252013] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.529 [2024-07-15 18:12:52.252030] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.529 [2024-07-15 18:12:52.252037] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.529 [2024-07-15 18:12:52.252043] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.529 [2024-07-15 18:12:52.252057] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.529 qpair failed and we were unable to recover it. 00:26:58.789 [2024-07-15 18:12:52.262009] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.789 [2024-07-15 18:12:52.262072] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.789 [2024-07-15 18:12:52.262088] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.789 [2024-07-15 18:12:52.262095] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.789 [2024-07-15 18:12:52.262102] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.789 [2024-07-15 18:12:52.262116] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.789 qpair failed and we were unable to recover it. 00:26:58.789 [2024-07-15 18:12:52.272027] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.789 [2024-07-15 18:12:52.272141] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.789 [2024-07-15 18:12:52.272159] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.789 [2024-07-15 18:12:52.272166] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.789 [2024-07-15 18:12:52.272172] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.789 [2024-07-15 18:12:52.272187] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.789 qpair failed and we were unable to recover it. 00:26:58.789 [2024-07-15 18:12:52.282033] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.789 [2024-07-15 18:12:52.282095] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.789 [2024-07-15 18:12:52.282110] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.789 [2024-07-15 18:12:52.282118] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.789 [2024-07-15 18:12:52.282125] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.789 [2024-07-15 18:12:52.282140] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.789 qpair failed and we were unable to recover it. 00:26:58.789 [2024-07-15 18:12:52.292079] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.789 [2024-07-15 18:12:52.292138] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.789 [2024-07-15 18:12:52.292153] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.789 [2024-07-15 18:12:52.292161] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.789 [2024-07-15 18:12:52.292167] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.789 [2024-07-15 18:12:52.292181] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.789 qpair failed and we were unable to recover it. 00:26:58.789 [2024-07-15 18:12:52.302114] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.789 [2024-07-15 18:12:52.302177] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.789 [2024-07-15 18:12:52.302194] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.789 [2024-07-15 18:12:52.302201] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.789 [2024-07-15 18:12:52.302207] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.789 [2024-07-15 18:12:52.302222] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.789 qpair failed and we were unable to recover it. 00:26:58.789 [2024-07-15 18:12:52.312146] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.789 [2024-07-15 18:12:52.312211] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.789 [2024-07-15 18:12:52.312232] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.789 [2024-07-15 18:12:52.312240] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.789 [2024-07-15 18:12:52.312246] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.789 [2024-07-15 18:12:52.312260] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.789 qpair failed and we were unable to recover it. 00:26:58.789 [2024-07-15 18:12:52.322170] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.789 [2024-07-15 18:12:52.322248] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.789 [2024-07-15 18:12:52.322266] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.789 [2024-07-15 18:12:52.322273] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.789 [2024-07-15 18:12:52.322283] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.789 [2024-07-15 18:12:52.322299] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.789 qpair failed and we were unable to recover it. 00:26:58.789 [2024-07-15 18:12:52.332176] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.789 [2024-07-15 18:12:52.332243] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.789 [2024-07-15 18:12:52.332259] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.789 [2024-07-15 18:12:52.332266] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.789 [2024-07-15 18:12:52.332272] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.789 [2024-07-15 18:12:52.332288] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.789 qpair failed and we were unable to recover it. 00:26:58.789 [2024-07-15 18:12:52.342238] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.789 [2024-07-15 18:12:52.342307] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.789 [2024-07-15 18:12:52.342322] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.789 [2024-07-15 18:12:52.342330] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.789 [2024-07-15 18:12:52.342336] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.789 [2024-07-15 18:12:52.342351] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.789 qpair failed and we were unable to recover it. 00:26:58.789 [2024-07-15 18:12:52.352253] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.789 [2024-07-15 18:12:52.352329] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.789 [2024-07-15 18:12:52.352345] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.789 [2024-07-15 18:12:52.352352] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.789 [2024-07-15 18:12:52.352358] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.790 [2024-07-15 18:12:52.352372] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.790 qpair failed and we were unable to recover it. 00:26:58.790 [2024-07-15 18:12:52.362270] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.790 [2024-07-15 18:12:52.362345] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.790 [2024-07-15 18:12:52.362361] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.790 [2024-07-15 18:12:52.362369] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.790 [2024-07-15 18:12:52.362375] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.790 [2024-07-15 18:12:52.362390] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.790 qpair failed and we were unable to recover it. 00:26:58.790 [2024-07-15 18:12:52.372285] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.790 [2024-07-15 18:12:52.372358] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.790 [2024-07-15 18:12:52.372374] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.790 [2024-07-15 18:12:52.372381] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.790 [2024-07-15 18:12:52.372388] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.790 [2024-07-15 18:12:52.372402] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.790 qpair failed and we were unable to recover it. 00:26:58.790 [2024-07-15 18:12:52.382339] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.790 [2024-07-15 18:12:52.382403] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.790 [2024-07-15 18:12:52.382418] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.790 [2024-07-15 18:12:52.382426] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.790 [2024-07-15 18:12:52.382432] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.790 [2024-07-15 18:12:52.382446] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.790 qpair failed and we were unable to recover it. 00:26:58.790 [2024-07-15 18:12:52.392289] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.790 [2024-07-15 18:12:52.392383] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.790 [2024-07-15 18:12:52.392400] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.790 [2024-07-15 18:12:52.392407] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.790 [2024-07-15 18:12:52.392415] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.790 [2024-07-15 18:12:52.392430] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.790 qpair failed and we were unable to recover it. 00:26:58.790 [2024-07-15 18:12:52.402381] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.790 [2024-07-15 18:12:52.402443] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.790 [2024-07-15 18:12:52.402460] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.790 [2024-07-15 18:12:52.402467] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.790 [2024-07-15 18:12:52.402473] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.790 [2024-07-15 18:12:52.402487] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.790 qpair failed and we were unable to recover it. 00:26:58.790 [2024-07-15 18:12:52.412453] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.790 [2024-07-15 18:12:52.412519] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.790 [2024-07-15 18:12:52.412535] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.790 [2024-07-15 18:12:52.412545] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.790 [2024-07-15 18:12:52.412551] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.790 [2024-07-15 18:12:52.412565] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.790 qpair failed and we were unable to recover it. 00:26:58.790 [2024-07-15 18:12:52.422461] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.790 [2024-07-15 18:12:52.422523] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.790 [2024-07-15 18:12:52.422539] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.790 [2024-07-15 18:12:52.422547] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.790 [2024-07-15 18:12:52.422553] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.790 [2024-07-15 18:12:52.422567] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.790 qpair failed and we were unable to recover it. 00:26:58.790 [2024-07-15 18:12:52.432465] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.790 [2024-07-15 18:12:52.432534] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.790 [2024-07-15 18:12:52.432550] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.790 [2024-07-15 18:12:52.432557] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.790 [2024-07-15 18:12:52.432563] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.790 [2024-07-15 18:12:52.432577] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.790 qpair failed and we were unable to recover it. 00:26:58.790 [2024-07-15 18:12:52.442517] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.790 [2024-07-15 18:12:52.442577] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.790 [2024-07-15 18:12:52.442594] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.790 [2024-07-15 18:12:52.442601] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.790 [2024-07-15 18:12:52.442607] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.790 [2024-07-15 18:12:52.442621] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.790 qpair failed and we were unable to recover it. 00:26:58.790 [2024-07-15 18:12:52.452509] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.790 [2024-07-15 18:12:52.452571] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.790 [2024-07-15 18:12:52.452586] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.790 [2024-07-15 18:12:52.452593] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.790 [2024-07-15 18:12:52.452600] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.790 [2024-07-15 18:12:52.452614] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.790 qpair failed and we were unable to recover it. 00:26:58.790 [2024-07-15 18:12:52.462541] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.790 [2024-07-15 18:12:52.462605] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.790 [2024-07-15 18:12:52.462621] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.790 [2024-07-15 18:12:52.462629] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.790 [2024-07-15 18:12:52.462636] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.790 [2024-07-15 18:12:52.462650] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.790 qpair failed and we were unable to recover it. 00:26:58.790 [2024-07-15 18:12:52.472580] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.790 [2024-07-15 18:12:52.472647] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.790 [2024-07-15 18:12:52.472662] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.790 [2024-07-15 18:12:52.472671] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.790 [2024-07-15 18:12:52.472677] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.790 [2024-07-15 18:12:52.472691] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.790 qpair failed and we were unable to recover it. 00:26:58.790 [2024-07-15 18:12:52.482612] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.790 [2024-07-15 18:12:52.482682] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.790 [2024-07-15 18:12:52.482699] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.790 [2024-07-15 18:12:52.482706] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.790 [2024-07-15 18:12:52.482712] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.790 [2024-07-15 18:12:52.482725] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.790 qpair failed and we were unable to recover it. 00:26:58.790 [2024-07-15 18:12:52.492645] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.790 [2024-07-15 18:12:52.492724] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.790 [2024-07-15 18:12:52.492740] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.790 [2024-07-15 18:12:52.492748] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.790 [2024-07-15 18:12:52.492754] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.791 [2024-07-15 18:12:52.492767] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.791 qpair failed and we were unable to recover it. 00:26:58.791 [2024-07-15 18:12:52.502666] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.791 [2024-07-15 18:12:52.502732] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.791 [2024-07-15 18:12:52.502748] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.791 [2024-07-15 18:12:52.502759] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.791 [2024-07-15 18:12:52.502765] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.791 [2024-07-15 18:12:52.502779] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.791 qpair failed and we were unable to recover it. 00:26:58.791 [2024-07-15 18:12:52.512605] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:58.791 [2024-07-15 18:12:52.512679] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:58.791 [2024-07-15 18:12:52.512695] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:58.791 [2024-07-15 18:12:52.512703] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:58.791 [2024-07-15 18:12:52.512708] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:58.791 [2024-07-15 18:12:52.512722] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:58.791 qpair failed and we were unable to recover it. 00:26:59.050 [2024-07-15 18:12:52.522704] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.050 [2024-07-15 18:12:52.522765] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.050 [2024-07-15 18:12:52.522781] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.050 [2024-07-15 18:12:52.522788] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.050 [2024-07-15 18:12:52.522794] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.050 [2024-07-15 18:12:52.522808] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.050 qpair failed and we were unable to recover it. 00:26:59.050 [2024-07-15 18:12:52.532756] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.050 [2024-07-15 18:12:52.532814] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.050 [2024-07-15 18:12:52.532829] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.050 [2024-07-15 18:12:52.532837] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.050 [2024-07-15 18:12:52.532843] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.050 [2024-07-15 18:12:52.532857] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.050 qpair failed and we were unable to recover it. 00:26:59.050 [2024-07-15 18:12:52.542746] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.050 [2024-07-15 18:12:52.542813] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.050 [2024-07-15 18:12:52.542828] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.050 [2024-07-15 18:12:52.542834] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.050 [2024-07-15 18:12:52.542841] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.050 [2024-07-15 18:12:52.542856] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.050 qpair failed and we were unable to recover it. 00:26:59.050 [2024-07-15 18:12:52.552802] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.050 [2024-07-15 18:12:52.552904] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.050 [2024-07-15 18:12:52.552920] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.050 [2024-07-15 18:12:52.552927] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.050 [2024-07-15 18:12:52.552933] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.050 [2024-07-15 18:12:52.552948] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.050 qpair failed and we were unable to recover it. 00:26:59.050 [2024-07-15 18:12:52.562768] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.050 [2024-07-15 18:12:52.562834] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.050 [2024-07-15 18:12:52.562850] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.050 [2024-07-15 18:12:52.562857] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.050 [2024-07-15 18:12:52.562863] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.050 [2024-07-15 18:12:52.562877] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.050 qpair failed and we were unable to recover it. 00:26:59.050 [2024-07-15 18:12:52.572881] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.050 [2024-07-15 18:12:52.572957] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.050 [2024-07-15 18:12:52.572973] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.050 [2024-07-15 18:12:52.572981] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.050 [2024-07-15 18:12:52.572987] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.050 [2024-07-15 18:12:52.573001] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.050 qpair failed and we were unable to recover it. 00:26:59.050 [2024-07-15 18:12:52.582982] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.050 [2024-07-15 18:12:52.583054] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.050 [2024-07-15 18:12:52.583073] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.050 [2024-07-15 18:12:52.583080] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.050 [2024-07-15 18:12:52.583086] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.050 [2024-07-15 18:12:52.583101] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.050 qpair failed and we were unable to recover it. 00:26:59.050 [2024-07-15 18:12:52.592978] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.050 [2024-07-15 18:12:52.593045] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.050 [2024-07-15 18:12:52.593061] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.050 [2024-07-15 18:12:52.593072] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.050 [2024-07-15 18:12:52.593078] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.050 [2024-07-15 18:12:52.593092] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.050 qpair failed and we were unable to recover it. 00:26:59.050 [2024-07-15 18:12:52.602992] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.050 [2024-07-15 18:12:52.603055] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.050 [2024-07-15 18:12:52.603072] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.050 [2024-07-15 18:12:52.603079] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.050 [2024-07-15 18:12:52.603086] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.050 [2024-07-15 18:12:52.603100] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.050 qpair failed and we were unable to recover it. 00:26:59.050 [2024-07-15 18:12:52.613041] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.050 [2024-07-15 18:12:52.613104] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.050 [2024-07-15 18:12:52.613119] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.050 [2024-07-15 18:12:52.613126] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.050 [2024-07-15 18:12:52.613132] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.050 [2024-07-15 18:12:52.613146] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.050 qpair failed and we were unable to recover it. 00:26:59.050 [2024-07-15 18:12:52.623008] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.050 [2024-07-15 18:12:52.623073] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.050 [2024-07-15 18:12:52.623090] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.050 [2024-07-15 18:12:52.623098] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.050 [2024-07-15 18:12:52.623104] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.050 [2024-07-15 18:12:52.623118] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.050 qpair failed and we were unable to recover it. 00:26:59.050 [2024-07-15 18:12:52.633041] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.050 [2024-07-15 18:12:52.633103] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.050 [2024-07-15 18:12:52.633119] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.050 [2024-07-15 18:12:52.633126] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.050 [2024-07-15 18:12:52.633133] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.050 [2024-07-15 18:12:52.633147] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.050 qpair failed and we were unable to recover it. 00:26:59.050 [2024-07-15 18:12:52.643059] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.050 [2024-07-15 18:12:52.643117] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.050 [2024-07-15 18:12:52.643132] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.050 [2024-07-15 18:12:52.643140] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.050 [2024-07-15 18:12:52.643146] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.050 [2024-07-15 18:12:52.643160] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.050 qpair failed and we were unable to recover it. 00:26:59.050 [2024-07-15 18:12:52.653069] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.050 [2024-07-15 18:12:52.653161] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.050 [2024-07-15 18:12:52.653178] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.050 [2024-07-15 18:12:52.653185] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.050 [2024-07-15 18:12:52.653191] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.050 [2024-07-15 18:12:52.653205] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.050 qpair failed and we were unable to recover it. 00:26:59.050 [2024-07-15 18:12:52.663126] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.051 [2024-07-15 18:12:52.663187] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.051 [2024-07-15 18:12:52.663202] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.051 [2024-07-15 18:12:52.663210] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.051 [2024-07-15 18:12:52.663216] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.051 [2024-07-15 18:12:52.663233] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.051 qpair failed and we were unable to recover it. 00:26:59.051 [2024-07-15 18:12:52.673170] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.051 [2024-07-15 18:12:52.673239] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.051 [2024-07-15 18:12:52.673255] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.051 [2024-07-15 18:12:52.673262] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.051 [2024-07-15 18:12:52.673269] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.051 [2024-07-15 18:12:52.673283] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.051 qpair failed and we were unable to recover it. 00:26:59.051 [2024-07-15 18:12:52.683168] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.051 [2024-07-15 18:12:52.683236] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.051 [2024-07-15 18:12:52.683255] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.051 [2024-07-15 18:12:52.683262] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.051 [2024-07-15 18:12:52.683268] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.051 [2024-07-15 18:12:52.683283] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.051 qpair failed and we were unable to recover it. 00:26:59.051 [2024-07-15 18:12:52.693195] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.051 [2024-07-15 18:12:52.693261] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.051 [2024-07-15 18:12:52.693277] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.051 [2024-07-15 18:12:52.693285] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.051 [2024-07-15 18:12:52.693291] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.051 [2024-07-15 18:12:52.693305] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.051 qpair failed and we were unable to recover it. 00:26:59.051 [2024-07-15 18:12:52.703234] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.051 [2024-07-15 18:12:52.703298] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.051 [2024-07-15 18:12:52.703313] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.051 [2024-07-15 18:12:52.703321] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.051 [2024-07-15 18:12:52.703327] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.051 [2024-07-15 18:12:52.703342] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.051 qpair failed and we were unable to recover it. 00:26:59.051 [2024-07-15 18:12:52.713275] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.051 [2024-07-15 18:12:52.713379] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.051 [2024-07-15 18:12:52.713395] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.051 [2024-07-15 18:12:52.713403] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.051 [2024-07-15 18:12:52.713410] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.051 [2024-07-15 18:12:52.713426] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.051 qpair failed and we were unable to recover it. 00:26:59.051 [2024-07-15 18:12:52.723302] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.051 [2024-07-15 18:12:52.723365] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.051 [2024-07-15 18:12:52.723380] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.051 [2024-07-15 18:12:52.723388] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.051 [2024-07-15 18:12:52.723394] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.051 [2024-07-15 18:12:52.723411] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.051 qpair failed and we were unable to recover it. 00:26:59.051 [2024-07-15 18:12:52.733316] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.051 [2024-07-15 18:12:52.733379] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.051 [2024-07-15 18:12:52.733397] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.051 [2024-07-15 18:12:52.733405] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.051 [2024-07-15 18:12:52.733411] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.051 [2024-07-15 18:12:52.733427] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.051 qpair failed and we were unable to recover it. 00:26:59.051 [2024-07-15 18:12:52.743364] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.051 [2024-07-15 18:12:52.743432] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.051 [2024-07-15 18:12:52.743448] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.051 [2024-07-15 18:12:52.743456] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.051 [2024-07-15 18:12:52.743462] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.051 [2024-07-15 18:12:52.743477] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.051 qpair failed and we were unable to recover it. 00:26:59.051 [2024-07-15 18:12:52.753360] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.051 [2024-07-15 18:12:52.753443] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.051 [2024-07-15 18:12:52.753460] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.051 [2024-07-15 18:12:52.753467] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.051 [2024-07-15 18:12:52.753474] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.051 [2024-07-15 18:12:52.753488] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.051 qpair failed and we were unable to recover it. 00:26:59.051 [2024-07-15 18:12:52.763337] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.051 [2024-07-15 18:12:52.763442] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.051 [2024-07-15 18:12:52.763458] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.051 [2024-07-15 18:12:52.763466] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.051 [2024-07-15 18:12:52.763473] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.051 [2024-07-15 18:12:52.763488] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.051 qpair failed and we were unable to recover it. 00:26:59.051 [2024-07-15 18:12:52.773414] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.051 [2024-07-15 18:12:52.773480] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.051 [2024-07-15 18:12:52.773500] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.051 [2024-07-15 18:12:52.773507] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.051 [2024-07-15 18:12:52.773513] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.051 [2024-07-15 18:12:52.773528] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.051 qpair failed and we were unable to recover it. 00:26:59.310 [2024-07-15 18:12:52.783479] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.310 [2024-07-15 18:12:52.783556] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.310 [2024-07-15 18:12:52.783572] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.310 [2024-07-15 18:12:52.783579] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.310 [2024-07-15 18:12:52.783585] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.310 [2024-07-15 18:12:52.783599] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.310 qpair failed and we were unable to recover it. 00:26:59.310 [2024-07-15 18:12:52.793507] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.310 [2024-07-15 18:12:52.793583] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.310 [2024-07-15 18:12:52.793599] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.310 [2024-07-15 18:12:52.793607] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.310 [2024-07-15 18:12:52.793613] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.310 [2024-07-15 18:12:52.793627] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.310 qpair failed and we were unable to recover it. 00:26:59.310 [2024-07-15 18:12:52.803452] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.310 [2024-07-15 18:12:52.803512] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.310 [2024-07-15 18:12:52.803528] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.310 [2024-07-15 18:12:52.803536] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.310 [2024-07-15 18:12:52.803542] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.310 [2024-07-15 18:12:52.803557] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.310 qpair failed and we were unable to recover it. 00:26:59.310 [2024-07-15 18:12:52.813512] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.310 [2024-07-15 18:12:52.813572] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.310 [2024-07-15 18:12:52.813587] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.310 [2024-07-15 18:12:52.813595] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.310 [2024-07-15 18:12:52.813601] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.310 [2024-07-15 18:12:52.813617] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.310 qpair failed and we were unable to recover it. 00:26:59.310 [2024-07-15 18:12:52.823576] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.310 [2024-07-15 18:12:52.823638] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.310 [2024-07-15 18:12:52.823654] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.310 [2024-07-15 18:12:52.823662] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.310 [2024-07-15 18:12:52.823668] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.310 [2024-07-15 18:12:52.823682] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.310 qpair failed and we were unable to recover it. 00:26:59.310 [2024-07-15 18:12:52.833610] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.310 [2024-07-15 18:12:52.833678] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.310 [2024-07-15 18:12:52.833693] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.310 [2024-07-15 18:12:52.833701] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.310 [2024-07-15 18:12:52.833707] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.310 [2024-07-15 18:12:52.833721] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.310 qpair failed and we were unable to recover it. 00:26:59.310 [2024-07-15 18:12:52.843656] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.310 [2024-07-15 18:12:52.843720] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.310 [2024-07-15 18:12:52.843735] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.310 [2024-07-15 18:12:52.843743] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.310 [2024-07-15 18:12:52.843749] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.310 [2024-07-15 18:12:52.843762] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.310 qpair failed and we were unable to recover it. 00:26:59.310 [2024-07-15 18:12:52.853643] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.310 [2024-07-15 18:12:52.853704] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.311 [2024-07-15 18:12:52.853720] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.311 [2024-07-15 18:12:52.853727] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.311 [2024-07-15 18:12:52.853734] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.311 [2024-07-15 18:12:52.853748] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.311 qpair failed and we were unable to recover it. 00:26:59.311 [2024-07-15 18:12:52.863701] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.311 [2024-07-15 18:12:52.863765] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.311 [2024-07-15 18:12:52.863784] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.311 [2024-07-15 18:12:52.863791] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.311 [2024-07-15 18:12:52.863797] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.311 [2024-07-15 18:12:52.863811] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.311 qpair failed and we were unable to recover it. 00:26:59.311 [2024-07-15 18:12:52.873722] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.311 [2024-07-15 18:12:52.873804] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.311 [2024-07-15 18:12:52.873819] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.311 [2024-07-15 18:12:52.873827] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.311 [2024-07-15 18:12:52.873833] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.311 [2024-07-15 18:12:52.873847] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.311 qpair failed and we were unable to recover it. 00:26:59.311 [2024-07-15 18:12:52.883727] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.311 [2024-07-15 18:12:52.883789] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.311 [2024-07-15 18:12:52.883807] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.311 [2024-07-15 18:12:52.883814] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.311 [2024-07-15 18:12:52.883821] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.311 [2024-07-15 18:12:52.883836] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.311 qpair failed and we were unable to recover it. 00:26:59.311 [2024-07-15 18:12:52.893709] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.311 [2024-07-15 18:12:52.893773] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.311 [2024-07-15 18:12:52.893789] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.311 [2024-07-15 18:12:52.893797] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.311 [2024-07-15 18:12:52.893803] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.311 [2024-07-15 18:12:52.893817] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.311 qpair failed and we were unable to recover it. 00:26:59.311 [2024-07-15 18:12:52.903742] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.311 [2024-07-15 18:12:52.903838] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.311 [2024-07-15 18:12:52.903854] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.311 [2024-07-15 18:12:52.903862] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.311 [2024-07-15 18:12:52.903868] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.311 [2024-07-15 18:12:52.903887] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.311 qpair failed and we were unable to recover it. 00:26:59.311 [2024-07-15 18:12:52.913775] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.311 [2024-07-15 18:12:52.913888] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.311 [2024-07-15 18:12:52.913905] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.311 [2024-07-15 18:12:52.913912] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.311 [2024-07-15 18:12:52.913919] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.311 [2024-07-15 18:12:52.913934] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.311 qpair failed and we were unable to recover it. 00:26:59.311 [2024-07-15 18:12:52.923835] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.311 [2024-07-15 18:12:52.923901] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.311 [2024-07-15 18:12:52.923917] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.311 [2024-07-15 18:12:52.923924] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.311 [2024-07-15 18:12:52.923930] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.311 [2024-07-15 18:12:52.923944] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.311 qpair failed and we were unable to recover it. 00:26:59.311 [2024-07-15 18:12:52.933903] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.311 [2024-07-15 18:12:52.933965] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.311 [2024-07-15 18:12:52.933980] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.311 [2024-07-15 18:12:52.933988] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.311 [2024-07-15 18:12:52.933995] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.311 [2024-07-15 18:12:52.934009] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.311 qpair failed and we were unable to recover it. 00:26:59.311 [2024-07-15 18:12:52.943907] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.311 [2024-07-15 18:12:52.943970] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.311 [2024-07-15 18:12:52.943986] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.311 [2024-07-15 18:12:52.943994] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.311 [2024-07-15 18:12:52.944000] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.311 [2024-07-15 18:12:52.944014] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.311 qpair failed and we were unable to recover it. 00:26:59.311 [2024-07-15 18:12:52.953905] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.311 [2024-07-15 18:12:52.953967] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.311 [2024-07-15 18:12:52.953985] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.311 [2024-07-15 18:12:52.953993] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.311 [2024-07-15 18:12:52.953999] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.311 [2024-07-15 18:12:52.954013] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.311 qpair failed and we were unable to recover it. 00:26:59.311 [2024-07-15 18:12:52.963961] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.311 [2024-07-15 18:12:52.964025] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.311 [2024-07-15 18:12:52.964041] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.311 [2024-07-15 18:12:52.964048] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.311 [2024-07-15 18:12:52.964054] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.311 [2024-07-15 18:12:52.964069] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.311 qpair failed and we were unable to recover it. 00:26:59.311 [2024-07-15 18:12:52.974023] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.311 [2024-07-15 18:12:52.974089] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.311 [2024-07-15 18:12:52.974106] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.311 [2024-07-15 18:12:52.974114] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.311 [2024-07-15 18:12:52.974120] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.311 [2024-07-15 18:12:52.974135] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.311 qpair failed and we were unable to recover it. 00:26:59.311 [2024-07-15 18:12:52.983963] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.311 [2024-07-15 18:12:52.984073] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.311 [2024-07-15 18:12:52.984089] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.311 [2024-07-15 18:12:52.984097] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.311 [2024-07-15 18:12:52.984104] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.311 [2024-07-15 18:12:52.984118] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.311 qpair failed and we were unable to recover it. 00:26:59.311 [2024-07-15 18:12:52.994064] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.311 [2024-07-15 18:12:52.994133] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.311 [2024-07-15 18:12:52.994149] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.311 [2024-07-15 18:12:52.994157] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.311 [2024-07-15 18:12:52.994166] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.311 [2024-07-15 18:12:52.994181] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.312 qpair failed and we were unable to recover it. 00:26:59.312 [2024-07-15 18:12:53.004082] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.312 [2024-07-15 18:12:53.004156] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.312 [2024-07-15 18:12:53.004173] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.312 [2024-07-15 18:12:53.004182] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.312 [2024-07-15 18:12:53.004187] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.312 [2024-07-15 18:12:53.004202] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.312 qpair failed and we were unable to recover it. 00:26:59.312 [2024-07-15 18:12:53.014086] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.312 [2024-07-15 18:12:53.014169] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.312 [2024-07-15 18:12:53.014185] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.312 [2024-07-15 18:12:53.014192] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.312 [2024-07-15 18:12:53.014198] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.312 [2024-07-15 18:12:53.014213] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.312 qpair failed and we were unable to recover it. 00:26:59.312 [2024-07-15 18:12:53.024056] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.312 [2024-07-15 18:12:53.024123] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.312 [2024-07-15 18:12:53.024139] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.312 [2024-07-15 18:12:53.024146] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.312 [2024-07-15 18:12:53.024152] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.312 [2024-07-15 18:12:53.024166] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.312 qpair failed and we were unable to recover it. 00:26:59.312 [2024-07-15 18:12:53.034073] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.312 [2024-07-15 18:12:53.034145] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.312 [2024-07-15 18:12:53.034161] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.312 [2024-07-15 18:12:53.034168] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.312 [2024-07-15 18:12:53.034175] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.312 [2024-07-15 18:12:53.034190] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.312 qpair failed and we were unable to recover it. 00:26:59.572 [2024-07-15 18:12:53.044152] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.572 [2024-07-15 18:12:53.044255] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.572 [2024-07-15 18:12:53.044273] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.572 [2024-07-15 18:12:53.044280] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.572 [2024-07-15 18:12:53.044286] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.572 [2024-07-15 18:12:53.044301] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.572 qpair failed and we were unable to recover it. 00:26:59.572 [2024-07-15 18:12:53.054139] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.572 [2024-07-15 18:12:53.054204] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.572 [2024-07-15 18:12:53.054222] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.572 [2024-07-15 18:12:53.054235] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.572 [2024-07-15 18:12:53.054241] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.572 [2024-07-15 18:12:53.054256] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.572 qpair failed and we were unable to recover it. 00:26:59.572 [2024-07-15 18:12:53.064235] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.572 [2024-07-15 18:12:53.064312] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.572 [2024-07-15 18:12:53.064329] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.572 [2024-07-15 18:12:53.064336] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.572 [2024-07-15 18:12:53.064343] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.572 [2024-07-15 18:12:53.064357] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.572 qpair failed and we were unable to recover it. 00:26:59.572 [2024-07-15 18:12:53.074203] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.572 [2024-07-15 18:12:53.074273] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.572 [2024-07-15 18:12:53.074289] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.572 [2024-07-15 18:12:53.074297] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.572 [2024-07-15 18:12:53.074304] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.572 [2024-07-15 18:12:53.074317] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.572 qpair failed and we were unable to recover it. 00:26:59.572 [2024-07-15 18:12:53.084270] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.572 [2024-07-15 18:12:53.084331] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.572 [2024-07-15 18:12:53.084348] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.572 [2024-07-15 18:12:53.084356] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.572 [2024-07-15 18:12:53.084370] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.572 [2024-07-15 18:12:53.084385] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.572 qpair failed and we were unable to recover it. 00:26:59.572 [2024-07-15 18:12:53.094392] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.572 [2024-07-15 18:12:53.094475] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.572 [2024-07-15 18:12:53.094492] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.572 [2024-07-15 18:12:53.094499] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.572 [2024-07-15 18:12:53.094505] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.572 [2024-07-15 18:12:53.094519] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.572 qpair failed and we were unable to recover it. 00:26:59.572 [2024-07-15 18:12:53.104296] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.572 [2024-07-15 18:12:53.104369] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.572 [2024-07-15 18:12:53.104385] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.572 [2024-07-15 18:12:53.104392] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.572 [2024-07-15 18:12:53.104398] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.572 [2024-07-15 18:12:53.104413] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.572 qpair failed and we were unable to recover it. 00:26:59.572 [2024-07-15 18:12:53.114325] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.572 [2024-07-15 18:12:53.114395] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.572 [2024-07-15 18:12:53.114410] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.572 [2024-07-15 18:12:53.114417] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.572 [2024-07-15 18:12:53.114423] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.572 [2024-07-15 18:12:53.114437] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.572 qpair failed and we were unable to recover it. 00:26:59.572 [2024-07-15 18:12:53.124346] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.572 [2024-07-15 18:12:53.124411] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.572 [2024-07-15 18:12:53.124427] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.572 [2024-07-15 18:12:53.124435] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.572 [2024-07-15 18:12:53.124441] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.572 [2024-07-15 18:12:53.124454] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.572 qpair failed and we were unable to recover it. 00:26:59.572 [2024-07-15 18:12:53.134428] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.572 [2024-07-15 18:12:53.134499] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.572 [2024-07-15 18:12:53.134515] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.572 [2024-07-15 18:12:53.134522] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.572 [2024-07-15 18:12:53.134528] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.572 [2024-07-15 18:12:53.134542] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.572 qpair failed and we were unable to recover it. 00:26:59.572 [2024-07-15 18:12:53.144406] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.572 [2024-07-15 18:12:53.144472] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.572 [2024-07-15 18:12:53.144488] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.572 [2024-07-15 18:12:53.144495] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.572 [2024-07-15 18:12:53.144501] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.573 [2024-07-15 18:12:53.144515] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.573 qpair failed and we were unable to recover it. 00:26:59.573 [2024-07-15 18:12:53.154512] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.573 [2024-07-15 18:12:53.154574] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.573 [2024-07-15 18:12:53.154590] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.573 [2024-07-15 18:12:53.154598] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.573 [2024-07-15 18:12:53.154604] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.573 [2024-07-15 18:12:53.154618] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.573 qpair failed and we were unable to recover it. 00:26:59.573 [2024-07-15 18:12:53.164452] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.573 [2024-07-15 18:12:53.164516] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.573 [2024-07-15 18:12:53.164532] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.573 [2024-07-15 18:12:53.164539] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.573 [2024-07-15 18:12:53.164545] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.573 [2024-07-15 18:12:53.164559] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.573 qpair failed and we were unable to recover it. 00:26:59.573 [2024-07-15 18:12:53.174547] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.573 [2024-07-15 18:12:53.174628] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.573 [2024-07-15 18:12:53.174644] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.573 [2024-07-15 18:12:53.174655] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.573 [2024-07-15 18:12:53.174661] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.573 [2024-07-15 18:12:53.174675] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.573 qpair failed and we were unable to recover it. 00:26:59.573 [2024-07-15 18:12:53.184538] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.573 [2024-07-15 18:12:53.184645] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.573 [2024-07-15 18:12:53.184660] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.573 [2024-07-15 18:12:53.184667] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.573 [2024-07-15 18:12:53.184674] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.573 [2024-07-15 18:12:53.184688] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.573 qpair failed and we were unable to recover it. 00:26:59.573 [2024-07-15 18:12:53.194624] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.573 [2024-07-15 18:12:53.194692] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.573 [2024-07-15 18:12:53.194708] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.573 [2024-07-15 18:12:53.194716] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.573 [2024-07-15 18:12:53.194722] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.573 [2024-07-15 18:12:53.194736] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.573 qpair failed and we were unable to recover it. 00:26:59.573 [2024-07-15 18:12:53.204623] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.573 [2024-07-15 18:12:53.204705] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.573 [2024-07-15 18:12:53.204720] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.573 [2024-07-15 18:12:53.204728] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.573 [2024-07-15 18:12:53.204734] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.573 [2024-07-15 18:12:53.204748] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.573 qpair failed and we were unable to recover it. 00:26:59.573 [2024-07-15 18:12:53.214644] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.573 [2024-07-15 18:12:53.214707] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.573 [2024-07-15 18:12:53.214723] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.573 [2024-07-15 18:12:53.214730] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.573 [2024-07-15 18:12:53.214736] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.573 [2024-07-15 18:12:53.214750] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.573 qpair failed and we were unable to recover it. 00:26:59.573 [2024-07-15 18:12:53.224671] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.573 [2024-07-15 18:12:53.224736] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.573 [2024-07-15 18:12:53.224752] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.573 [2024-07-15 18:12:53.224759] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.573 [2024-07-15 18:12:53.224766] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.573 [2024-07-15 18:12:53.224781] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.573 qpair failed and we were unable to recover it. 00:26:59.573 [2024-07-15 18:12:53.234709] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.573 [2024-07-15 18:12:53.234783] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.573 [2024-07-15 18:12:53.234798] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.573 [2024-07-15 18:12:53.234806] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.573 [2024-07-15 18:12:53.234812] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.573 [2024-07-15 18:12:53.234826] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.573 qpair failed and we were unable to recover it. 00:26:59.573 [2024-07-15 18:12:53.244790] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.573 [2024-07-15 18:12:53.244854] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.573 [2024-07-15 18:12:53.244869] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.573 [2024-07-15 18:12:53.244877] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.573 [2024-07-15 18:12:53.244882] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.573 [2024-07-15 18:12:53.244897] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.573 qpair failed and we were unable to recover it. 00:26:59.573 [2024-07-15 18:12:53.254771] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.573 [2024-07-15 18:12:53.254833] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.573 [2024-07-15 18:12:53.254848] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.573 [2024-07-15 18:12:53.254856] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.573 [2024-07-15 18:12:53.254862] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.573 [2024-07-15 18:12:53.254877] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.573 qpair failed and we were unable to recover it. 00:26:59.573 [2024-07-15 18:12:53.264804] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.573 [2024-07-15 18:12:53.264878] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.573 [2024-07-15 18:12:53.264894] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.573 [2024-07-15 18:12:53.264904] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.573 [2024-07-15 18:12:53.264910] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.573 [2024-07-15 18:12:53.264924] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.573 qpair failed and we were unable to recover it. 00:26:59.573 [2024-07-15 18:12:53.274823] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.573 [2024-07-15 18:12:53.274884] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.573 [2024-07-15 18:12:53.274900] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.573 [2024-07-15 18:12:53.274908] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.573 [2024-07-15 18:12:53.274914] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.573 [2024-07-15 18:12:53.274928] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.573 qpair failed and we were unable to recover it. 00:26:59.574 [2024-07-15 18:12:53.284834] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.574 [2024-07-15 18:12:53.284897] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.574 [2024-07-15 18:12:53.284912] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.574 [2024-07-15 18:12:53.284920] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.574 [2024-07-15 18:12:53.284927] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.574 [2024-07-15 18:12:53.284940] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.574 qpair failed and we were unable to recover it. 00:26:59.574 [2024-07-15 18:12:53.294861] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.574 [2024-07-15 18:12:53.294929] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.574 [2024-07-15 18:12:53.294944] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.574 [2024-07-15 18:12:53.294951] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.574 [2024-07-15 18:12:53.294958] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.574 [2024-07-15 18:12:53.294972] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.574 qpair failed and we were unable to recover it. 00:26:59.833 [2024-07-15 18:12:53.304843] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.833 [2024-07-15 18:12:53.304906] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.833 [2024-07-15 18:12:53.304922] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.833 [2024-07-15 18:12:53.304929] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.833 [2024-07-15 18:12:53.304935] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.833 [2024-07-15 18:12:53.304949] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.833 qpair failed and we were unable to recover it. 00:26:59.833 [2024-07-15 18:12:53.314932] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.833 [2024-07-15 18:12:53.314993] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.833 [2024-07-15 18:12:53.315009] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.833 [2024-07-15 18:12:53.315017] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.833 [2024-07-15 18:12:53.315023] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.833 [2024-07-15 18:12:53.315038] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.833 qpair failed and we were unable to recover it. 00:26:59.833 [2024-07-15 18:12:53.324967] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.833 [2024-07-15 18:12:53.325034] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.833 [2024-07-15 18:12:53.325050] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.833 [2024-07-15 18:12:53.325057] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.833 [2024-07-15 18:12:53.325063] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.833 [2024-07-15 18:12:53.325077] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.833 qpair failed and we were unable to recover it. 00:26:59.833 [2024-07-15 18:12:53.335002] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.833 [2024-07-15 18:12:53.335065] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.834 [2024-07-15 18:12:53.335081] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.834 [2024-07-15 18:12:53.335088] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.834 [2024-07-15 18:12:53.335094] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.834 [2024-07-15 18:12:53.335108] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.834 qpair failed and we were unable to recover it. 00:26:59.834 [2024-07-15 18:12:53.345009] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.834 [2024-07-15 18:12:53.345071] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.834 [2024-07-15 18:12:53.345087] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.834 [2024-07-15 18:12:53.345094] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.834 [2024-07-15 18:12:53.345100] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.834 [2024-07-15 18:12:53.345114] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.834 qpair failed and we were unable to recover it. 00:26:59.834 [2024-07-15 18:12:53.355062] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.834 [2024-07-15 18:12:53.355126] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.834 [2024-07-15 18:12:53.355141] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.834 [2024-07-15 18:12:53.355152] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.834 [2024-07-15 18:12:53.355158] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.834 [2024-07-15 18:12:53.355172] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.834 qpair failed and we were unable to recover it. 00:26:59.834 [2024-07-15 18:12:53.365003] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.834 [2024-07-15 18:12:53.365068] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.834 [2024-07-15 18:12:53.365085] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.834 [2024-07-15 18:12:53.365092] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.834 [2024-07-15 18:12:53.365098] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.834 [2024-07-15 18:12:53.365113] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.834 qpair failed and we were unable to recover it. 00:26:59.834 [2024-07-15 18:12:53.375073] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.834 [2024-07-15 18:12:53.375135] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.834 [2024-07-15 18:12:53.375150] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.834 [2024-07-15 18:12:53.375158] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.834 [2024-07-15 18:12:53.375164] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.834 [2024-07-15 18:12:53.375178] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.834 qpair failed and we were unable to recover it. 00:26:59.834 [2024-07-15 18:12:53.385124] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.834 [2024-07-15 18:12:53.385187] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.834 [2024-07-15 18:12:53.385203] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.834 [2024-07-15 18:12:53.385210] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.834 [2024-07-15 18:12:53.385216] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.834 [2024-07-15 18:12:53.385242] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.834 qpair failed and we were unable to recover it. 00:26:59.834 [2024-07-15 18:12:53.395152] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.834 [2024-07-15 18:12:53.395230] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.834 [2024-07-15 18:12:53.395246] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.834 [2024-07-15 18:12:53.395254] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.834 [2024-07-15 18:12:53.395260] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.834 [2024-07-15 18:12:53.395275] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.834 qpair failed and we were unable to recover it. 00:26:59.834 [2024-07-15 18:12:53.405193] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.834 [2024-07-15 18:12:53.405261] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.834 [2024-07-15 18:12:53.405278] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.834 [2024-07-15 18:12:53.405285] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.834 [2024-07-15 18:12:53.405292] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.834 [2024-07-15 18:12:53.405306] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.834 qpair failed and we were unable to recover it. 00:26:59.834 [2024-07-15 18:12:53.415231] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.834 [2024-07-15 18:12:53.415295] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.834 [2024-07-15 18:12:53.415311] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.834 [2024-07-15 18:12:53.415318] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.834 [2024-07-15 18:12:53.415325] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.834 [2024-07-15 18:12:53.415339] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.834 qpair failed and we were unable to recover it. 00:26:59.834 [2024-07-15 18:12:53.425263] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.834 [2024-07-15 18:12:53.425328] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.834 [2024-07-15 18:12:53.425345] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.834 [2024-07-15 18:12:53.425352] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.834 [2024-07-15 18:12:53.425359] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.834 [2024-07-15 18:12:53.425373] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.834 qpair failed and we were unable to recover it. 00:26:59.834 [2024-07-15 18:12:53.435290] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.834 [2024-07-15 18:12:53.435359] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.834 [2024-07-15 18:12:53.435375] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.834 [2024-07-15 18:12:53.435382] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.834 [2024-07-15 18:12:53.435388] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.834 [2024-07-15 18:12:53.435403] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.834 qpair failed and we were unable to recover it. 00:26:59.834 [2024-07-15 18:12:53.445288] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.834 [2024-07-15 18:12:53.445373] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.834 [2024-07-15 18:12:53.445394] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.834 [2024-07-15 18:12:53.445402] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.834 [2024-07-15 18:12:53.445408] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.834 [2024-07-15 18:12:53.445423] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.834 qpair failed and we were unable to recover it. 00:26:59.834 [2024-07-15 18:12:53.455314] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.834 [2024-07-15 18:12:53.455381] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.834 [2024-07-15 18:12:53.455396] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.834 [2024-07-15 18:12:53.455403] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.834 [2024-07-15 18:12:53.455409] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.834 [2024-07-15 18:12:53.455423] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.834 qpair failed and we were unable to recover it. 00:26:59.834 [2024-07-15 18:12:53.465349] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.834 [2024-07-15 18:12:53.465415] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.834 [2024-07-15 18:12:53.465432] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.834 [2024-07-15 18:12:53.465440] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.834 [2024-07-15 18:12:53.465446] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.834 [2024-07-15 18:12:53.465461] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.834 qpair failed and we were unable to recover it. 00:26:59.834 [2024-07-15 18:12:53.475394] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.834 [2024-07-15 18:12:53.475464] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.834 [2024-07-15 18:12:53.475480] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.834 [2024-07-15 18:12:53.475488] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.834 [2024-07-15 18:12:53.475494] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.834 [2024-07-15 18:12:53.475508] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.834 qpair failed and we were unable to recover it. 00:26:59.834 [2024-07-15 18:12:53.485425] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.834 [2024-07-15 18:12:53.485484] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.834 [2024-07-15 18:12:53.485500] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.834 [2024-07-15 18:12:53.485508] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.834 [2024-07-15 18:12:53.485514] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.834 [2024-07-15 18:12:53.485528] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.834 qpair failed and we were unable to recover it. 00:26:59.834 [2024-07-15 18:12:53.495454] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.834 [2024-07-15 18:12:53.495518] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.834 [2024-07-15 18:12:53.495534] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.834 [2024-07-15 18:12:53.495541] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.834 [2024-07-15 18:12:53.495547] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.834 [2024-07-15 18:12:53.495561] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.834 qpair failed and we were unable to recover it. 00:26:59.835 [2024-07-15 18:12:53.505469] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.835 [2024-07-15 18:12:53.505534] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.835 [2024-07-15 18:12:53.505549] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.835 [2024-07-15 18:12:53.505556] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.835 [2024-07-15 18:12:53.505562] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.835 [2024-07-15 18:12:53.505576] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.835 qpair failed and we were unable to recover it. 00:26:59.835 [2024-07-15 18:12:53.515509] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.835 [2024-07-15 18:12:53.515575] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.835 [2024-07-15 18:12:53.515590] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.835 [2024-07-15 18:12:53.515598] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.835 [2024-07-15 18:12:53.515604] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.835 [2024-07-15 18:12:53.515618] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.835 qpair failed and we were unable to recover it. 00:26:59.835 [2024-07-15 18:12:53.525532] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.835 [2024-07-15 18:12:53.525591] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.835 [2024-07-15 18:12:53.525606] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.835 [2024-07-15 18:12:53.525614] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.835 [2024-07-15 18:12:53.525620] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.835 [2024-07-15 18:12:53.525635] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.835 qpair failed and we were unable to recover it. 00:26:59.835 [2024-07-15 18:12:53.535613] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.835 [2024-07-15 18:12:53.535674] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.835 [2024-07-15 18:12:53.535693] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.835 [2024-07-15 18:12:53.535700] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.835 [2024-07-15 18:12:53.535707] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.835 [2024-07-15 18:12:53.535722] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.835 qpair failed and we were unable to recover it. 00:26:59.835 [2024-07-15 18:12:53.545630] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.835 [2024-07-15 18:12:53.545693] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.835 [2024-07-15 18:12:53.545710] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.835 [2024-07-15 18:12:53.545717] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.835 [2024-07-15 18:12:53.545724] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.835 [2024-07-15 18:12:53.545738] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.835 qpair failed and we were unable to recover it. 00:26:59.835 [2024-07-15 18:12:53.555634] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:26:59.835 [2024-07-15 18:12:53.555705] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:26:59.835 [2024-07-15 18:12:53.555720] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:26:59.835 [2024-07-15 18:12:53.555728] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:26:59.835 [2024-07-15 18:12:53.555734] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:26:59.835 [2024-07-15 18:12:53.555748] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:26:59.835 qpair failed and we were unable to recover it. 00:27:00.094 [2024-07-15 18:12:53.565633] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.094 [2024-07-15 18:12:53.565698] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.094 [2024-07-15 18:12:53.565714] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.094 [2024-07-15 18:12:53.565721] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.094 [2024-07-15 18:12:53.565727] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.094 [2024-07-15 18:12:53.565741] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.094 qpair failed and we were unable to recover it. 00:27:00.094 [2024-07-15 18:12:53.575650] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.094 [2024-07-15 18:12:53.575711] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.094 [2024-07-15 18:12:53.575727] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.094 [2024-07-15 18:12:53.575734] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.094 [2024-07-15 18:12:53.575740] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.094 [2024-07-15 18:12:53.575757] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.094 qpair failed and we were unable to recover it. 00:27:00.094 [2024-07-15 18:12:53.585715] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.094 [2024-07-15 18:12:53.585779] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.094 [2024-07-15 18:12:53.585795] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.094 [2024-07-15 18:12:53.585802] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.094 [2024-07-15 18:12:53.585808] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.094 [2024-07-15 18:12:53.585823] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.094 qpair failed and we were unable to recover it. 00:27:00.094 [2024-07-15 18:12:53.595736] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.094 [2024-07-15 18:12:53.595842] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.094 [2024-07-15 18:12:53.595859] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.094 [2024-07-15 18:12:53.595866] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.094 [2024-07-15 18:12:53.595873] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.094 [2024-07-15 18:12:53.595889] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.094 qpair failed and we were unable to recover it. 00:27:00.094 [2024-07-15 18:12:53.605732] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.094 [2024-07-15 18:12:53.605796] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.095 [2024-07-15 18:12:53.605812] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.095 [2024-07-15 18:12:53.605819] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.095 [2024-07-15 18:12:53.605826] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.095 [2024-07-15 18:12:53.605840] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.095 qpair failed and we were unable to recover it. 00:27:00.095 [2024-07-15 18:12:53.615771] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.095 [2024-07-15 18:12:53.615838] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.095 [2024-07-15 18:12:53.615854] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.095 [2024-07-15 18:12:53.615861] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.095 [2024-07-15 18:12:53.615867] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.095 [2024-07-15 18:12:53.615882] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.095 qpair failed and we were unable to recover it. 00:27:00.095 [2024-07-15 18:12:53.625802] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.095 [2024-07-15 18:12:53.625867] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.095 [2024-07-15 18:12:53.625885] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.095 [2024-07-15 18:12:53.625892] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.095 [2024-07-15 18:12:53.625899] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.095 [2024-07-15 18:12:53.625913] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.095 qpair failed and we were unable to recover it. 00:27:00.095 [2024-07-15 18:12:53.635820] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.095 [2024-07-15 18:12:53.635886] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.095 [2024-07-15 18:12:53.635901] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.095 [2024-07-15 18:12:53.635909] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.095 [2024-07-15 18:12:53.635915] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.095 [2024-07-15 18:12:53.635928] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.095 qpair failed and we were unable to recover it. 00:27:00.095 [2024-07-15 18:12:53.645911] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.095 [2024-07-15 18:12:53.645977] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.095 [2024-07-15 18:12:53.645993] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.095 [2024-07-15 18:12:53.646000] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.095 [2024-07-15 18:12:53.646006] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.095 [2024-07-15 18:12:53.646021] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.095 qpair failed and we were unable to recover it. 00:27:00.095 [2024-07-15 18:12:53.655875] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.095 [2024-07-15 18:12:53.655936] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.095 [2024-07-15 18:12:53.655952] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.095 [2024-07-15 18:12:53.655959] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.095 [2024-07-15 18:12:53.655966] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.095 [2024-07-15 18:12:53.655980] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.095 qpair failed and we were unable to recover it. 00:27:00.095 [2024-07-15 18:12:53.665916] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.095 [2024-07-15 18:12:53.665977] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.095 [2024-07-15 18:12:53.665992] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.095 [2024-07-15 18:12:53.666000] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.095 [2024-07-15 18:12:53.666006] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.095 [2024-07-15 18:12:53.666023] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.095 qpair failed and we were unable to recover it. 00:27:00.095 [2024-07-15 18:12:53.675873] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.095 [2024-07-15 18:12:53.675978] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.095 [2024-07-15 18:12:53.675993] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.095 [2024-07-15 18:12:53.676001] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.095 [2024-07-15 18:12:53.676008] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.095 [2024-07-15 18:12:53.676022] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.095 qpair failed and we were unable to recover it. 00:27:00.095 [2024-07-15 18:12:53.685955] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.095 [2024-07-15 18:12:53.686014] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.095 [2024-07-15 18:12:53.686030] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.095 [2024-07-15 18:12:53.686038] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.095 [2024-07-15 18:12:53.686044] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.095 [2024-07-15 18:12:53.686058] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.095 qpair failed and we were unable to recover it. 00:27:00.095 [2024-07-15 18:12:53.695983] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.095 [2024-07-15 18:12:53.696042] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.095 [2024-07-15 18:12:53.696057] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.095 [2024-07-15 18:12:53.696064] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.095 [2024-07-15 18:12:53.696070] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.095 [2024-07-15 18:12:53.696084] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.095 qpair failed and we were unable to recover it. 00:27:00.095 [2024-07-15 18:12:53.706035] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.095 [2024-07-15 18:12:53.706101] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.095 [2024-07-15 18:12:53.706116] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.095 [2024-07-15 18:12:53.706123] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.095 [2024-07-15 18:12:53.706129] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.095 [2024-07-15 18:12:53.706143] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.095 qpair failed and we were unable to recover it. 00:27:00.095 [2024-07-15 18:12:53.716116] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.095 [2024-07-15 18:12:53.716207] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.095 [2024-07-15 18:12:53.716229] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.095 [2024-07-15 18:12:53.716237] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.095 [2024-07-15 18:12:53.716243] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.095 [2024-07-15 18:12:53.716258] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.095 qpair failed and we were unable to recover it. 00:27:00.095 [2024-07-15 18:12:53.726117] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.095 [2024-07-15 18:12:53.726232] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.095 [2024-07-15 18:12:53.726248] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.095 [2024-07-15 18:12:53.726255] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.095 [2024-07-15 18:12:53.726263] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.095 [2024-07-15 18:12:53.726277] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.095 qpair failed and we were unable to recover it. 00:27:00.095 [2024-07-15 18:12:53.736150] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.096 [2024-07-15 18:12:53.736238] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.096 [2024-07-15 18:12:53.736256] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.096 [2024-07-15 18:12:53.736264] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.096 [2024-07-15 18:12:53.736270] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.096 [2024-07-15 18:12:53.736285] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.096 qpair failed and we were unable to recover it. 00:27:00.096 [2024-07-15 18:12:53.746162] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.096 [2024-07-15 18:12:53.746234] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.096 [2024-07-15 18:12:53.746249] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.096 [2024-07-15 18:12:53.746257] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.096 [2024-07-15 18:12:53.746263] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.096 [2024-07-15 18:12:53.746278] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.096 qpair failed and we were unable to recover it. 00:27:00.096 [2024-07-15 18:12:53.756153] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.096 [2024-07-15 18:12:53.756219] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.096 [2024-07-15 18:12:53.756237] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.096 [2024-07-15 18:12:53.756245] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.096 [2024-07-15 18:12:53.756255] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.096 [2024-07-15 18:12:53.756269] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.096 qpair failed and we were unable to recover it. 00:27:00.096 [2024-07-15 18:12:53.766209] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.096 [2024-07-15 18:12:53.766274] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.096 [2024-07-15 18:12:53.766290] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.096 [2024-07-15 18:12:53.766297] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.096 [2024-07-15 18:12:53.766303] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.096 [2024-07-15 18:12:53.766318] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.096 qpair failed and we were unable to recover it. 00:27:00.096 [2024-07-15 18:12:53.776276] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.096 [2024-07-15 18:12:53.776382] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.096 [2024-07-15 18:12:53.776400] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.096 [2024-07-15 18:12:53.776407] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.096 [2024-07-15 18:12:53.776414] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.096 [2024-07-15 18:12:53.776428] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.096 qpair failed and we were unable to recover it. 00:27:00.096 [2024-07-15 18:12:53.786279] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.096 [2024-07-15 18:12:53.786340] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.096 [2024-07-15 18:12:53.786355] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.096 [2024-07-15 18:12:53.786363] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.096 [2024-07-15 18:12:53.786369] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.096 [2024-07-15 18:12:53.786384] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.096 qpair failed and we were unable to recover it. 00:27:00.096 [2024-07-15 18:12:53.796294] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.096 [2024-07-15 18:12:53.796366] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.096 [2024-07-15 18:12:53.796383] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.096 [2024-07-15 18:12:53.796390] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.096 [2024-07-15 18:12:53.796397] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.096 [2024-07-15 18:12:53.796411] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.096 qpair failed and we were unable to recover it. 00:27:00.096 [2024-07-15 18:12:53.806336] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.096 [2024-07-15 18:12:53.806415] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.096 [2024-07-15 18:12:53.806431] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.096 [2024-07-15 18:12:53.806438] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.096 [2024-07-15 18:12:53.806444] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.096 [2024-07-15 18:12:53.806458] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.096 qpair failed and we were unable to recover it. 00:27:00.096 [2024-07-15 18:12:53.816389] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.096 [2024-07-15 18:12:53.816449] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.096 [2024-07-15 18:12:53.816465] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.096 [2024-07-15 18:12:53.816472] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.096 [2024-07-15 18:12:53.816478] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.096 [2024-07-15 18:12:53.816493] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.096 qpair failed and we were unable to recover it. 00:27:00.356 [2024-07-15 18:12:53.826392] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.356 [2024-07-15 18:12:53.826456] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.356 [2024-07-15 18:12:53.826472] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.356 [2024-07-15 18:12:53.826479] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.356 [2024-07-15 18:12:53.826485] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.356 [2024-07-15 18:12:53.826499] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.356 qpair failed and we were unable to recover it. 00:27:00.356 [2024-07-15 18:12:53.836412] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.356 [2024-07-15 18:12:53.836484] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.356 [2024-07-15 18:12:53.836499] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.356 [2024-07-15 18:12:53.836507] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.356 [2024-07-15 18:12:53.836513] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.356 [2024-07-15 18:12:53.836527] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.356 qpair failed and we were unable to recover it. 00:27:00.356 [2024-07-15 18:12:53.846434] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.356 [2024-07-15 18:12:53.846501] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.356 [2024-07-15 18:12:53.846517] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.356 [2024-07-15 18:12:53.846524] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.356 [2024-07-15 18:12:53.846536] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.356 [2024-07-15 18:12:53.846550] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.356 qpair failed and we were unable to recover it. 00:27:00.356 [2024-07-15 18:12:53.856509] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.356 [2024-07-15 18:12:53.856624] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.356 [2024-07-15 18:12:53.856641] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.356 [2024-07-15 18:12:53.856649] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.356 [2024-07-15 18:12:53.856656] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.356 [2024-07-15 18:12:53.856670] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.356 qpair failed and we were unable to recover it. 00:27:00.356 [2024-07-15 18:12:53.866504] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.356 [2024-07-15 18:12:53.866578] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.356 [2024-07-15 18:12:53.866594] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.356 [2024-07-15 18:12:53.866602] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.356 [2024-07-15 18:12:53.866608] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.356 [2024-07-15 18:12:53.866622] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.356 qpair failed and we were unable to recover it. 00:27:00.356 [2024-07-15 18:12:53.876535] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.356 [2024-07-15 18:12:53.876613] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.356 [2024-07-15 18:12:53.876629] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.356 [2024-07-15 18:12:53.876636] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.356 [2024-07-15 18:12:53.876643] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.356 [2024-07-15 18:12:53.876657] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.356 qpair failed and we were unable to recover it. 00:27:00.356 [2024-07-15 18:12:53.886555] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.356 [2024-07-15 18:12:53.886659] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.356 [2024-07-15 18:12:53.886674] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.356 [2024-07-15 18:12:53.886681] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.356 [2024-07-15 18:12:53.886688] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.356 [2024-07-15 18:12:53.886703] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.356 qpair failed and we were unable to recover it. 00:27:00.356 [2024-07-15 18:12:53.896590] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.356 [2024-07-15 18:12:53.896693] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.356 [2024-07-15 18:12:53.896709] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.356 [2024-07-15 18:12:53.896717] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.356 [2024-07-15 18:12:53.896724] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.356 [2024-07-15 18:12:53.896738] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.356 qpair failed and we were unable to recover it. 00:27:00.356 [2024-07-15 18:12:53.906631] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.356 [2024-07-15 18:12:53.906694] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.356 [2024-07-15 18:12:53.906710] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.356 [2024-07-15 18:12:53.906717] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.356 [2024-07-15 18:12:53.906723] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.356 [2024-07-15 18:12:53.906737] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.356 qpair failed and we were unable to recover it. 00:27:00.356 [2024-07-15 18:12:53.916659] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.356 [2024-07-15 18:12:53.916734] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.356 [2024-07-15 18:12:53.916749] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.356 [2024-07-15 18:12:53.916757] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.356 [2024-07-15 18:12:53.916763] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.356 [2024-07-15 18:12:53.916777] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.356 qpair failed and we were unable to recover it. 00:27:00.356 [2024-07-15 18:12:53.926687] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.356 [2024-07-15 18:12:53.926756] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.356 [2024-07-15 18:12:53.926772] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.356 [2024-07-15 18:12:53.926779] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.356 [2024-07-15 18:12:53.926785] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.356 [2024-07-15 18:12:53.926799] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.356 qpair failed and we were unable to recover it. 00:27:00.356 [2024-07-15 18:12:53.936764] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.356 [2024-07-15 18:12:53.936829] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.356 [2024-07-15 18:12:53.936846] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.356 [2024-07-15 18:12:53.936854] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.356 [2024-07-15 18:12:53.936864] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.356 [2024-07-15 18:12:53.936879] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.356 qpair failed and we were unable to recover it. 00:27:00.356 [2024-07-15 18:12:53.946752] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.356 [2024-07-15 18:12:53.946815] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.356 [2024-07-15 18:12:53.946830] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.356 [2024-07-15 18:12:53.946838] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.357 [2024-07-15 18:12:53.946844] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.357 [2024-07-15 18:12:53.946858] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.357 qpair failed and we were unable to recover it. 00:27:00.357 [2024-07-15 18:12:53.956732] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.357 [2024-07-15 18:12:53.956814] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.357 [2024-07-15 18:12:53.956830] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.357 [2024-07-15 18:12:53.956837] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.357 [2024-07-15 18:12:53.956843] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.357 [2024-07-15 18:12:53.956857] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.357 qpair failed and we were unable to recover it. 00:27:00.357 [2024-07-15 18:12:53.966804] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.357 [2024-07-15 18:12:53.966918] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.357 [2024-07-15 18:12:53.966936] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.357 [2024-07-15 18:12:53.966943] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.357 [2024-07-15 18:12:53.966950] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.357 [2024-07-15 18:12:53.966966] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.357 qpair failed and we were unable to recover it. 00:27:00.357 [2024-07-15 18:12:53.976809] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.357 [2024-07-15 18:12:53.976871] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.357 [2024-07-15 18:12:53.976886] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.357 [2024-07-15 18:12:53.976894] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.357 [2024-07-15 18:12:53.976901] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.357 [2024-07-15 18:12:53.976915] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.357 qpair failed and we were unable to recover it. 00:27:00.357 [2024-07-15 18:12:53.986851] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.357 [2024-07-15 18:12:53.986919] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.357 [2024-07-15 18:12:53.986934] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.357 [2024-07-15 18:12:53.986942] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.357 [2024-07-15 18:12:53.986948] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.357 [2024-07-15 18:12:53.986963] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.357 qpair failed and we were unable to recover it. 00:27:00.357 [2024-07-15 18:12:53.996892] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.357 [2024-07-15 18:12:53.996998] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.357 [2024-07-15 18:12:53.997013] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.357 [2024-07-15 18:12:53.997020] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.357 [2024-07-15 18:12:53.997027] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.357 [2024-07-15 18:12:53.997043] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.357 qpair failed and we were unable to recover it. 00:27:00.357 [2024-07-15 18:12:54.006978] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.357 [2024-07-15 18:12:54.007080] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.357 [2024-07-15 18:12:54.007096] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.357 [2024-07-15 18:12:54.007104] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.357 [2024-07-15 18:12:54.007111] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.357 [2024-07-15 18:12:54.007126] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.357 qpair failed and we were unable to recover it. 00:27:00.357 [2024-07-15 18:12:54.016981] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.357 [2024-07-15 18:12:54.017043] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.357 [2024-07-15 18:12:54.017059] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.357 [2024-07-15 18:12:54.017066] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.357 [2024-07-15 18:12:54.017072] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.357 [2024-07-15 18:12:54.017086] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.357 qpair failed and we were unable to recover it. 00:27:00.357 [2024-07-15 18:12:54.026999] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.357 [2024-07-15 18:12:54.027065] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.357 [2024-07-15 18:12:54.027080] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.357 [2024-07-15 18:12:54.027090] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.357 [2024-07-15 18:12:54.027096] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.357 [2024-07-15 18:12:54.027111] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.357 qpair failed and we were unable to recover it. 00:27:00.357 [2024-07-15 18:12:54.036997] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.357 [2024-07-15 18:12:54.037061] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.357 [2024-07-15 18:12:54.037076] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.357 [2024-07-15 18:12:54.037084] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.357 [2024-07-15 18:12:54.037090] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.357 [2024-07-15 18:12:54.037103] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.357 qpair failed and we were unable to recover it. 00:27:00.357 [2024-07-15 18:12:54.047028] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.357 [2024-07-15 18:12:54.047092] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.357 [2024-07-15 18:12:54.047109] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.357 [2024-07-15 18:12:54.047117] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.357 [2024-07-15 18:12:54.047124] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.357 [2024-07-15 18:12:54.047138] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.357 qpair failed and we were unable to recover it. 00:27:00.357 [2024-07-15 18:12:54.057046] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.357 [2024-07-15 18:12:54.057116] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.357 [2024-07-15 18:12:54.057132] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.357 [2024-07-15 18:12:54.057139] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.357 [2024-07-15 18:12:54.057146] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.357 [2024-07-15 18:12:54.057160] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.357 qpair failed and we were unable to recover it. 00:27:00.357 [2024-07-15 18:12:54.067088] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.357 [2024-07-15 18:12:54.067152] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.357 [2024-07-15 18:12:54.067168] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.357 [2024-07-15 18:12:54.067175] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.357 [2024-07-15 18:12:54.067181] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.357 [2024-07-15 18:12:54.067195] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.357 qpair failed and we were unable to recover it. 00:27:00.357 [2024-07-15 18:12:54.077113] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.357 [2024-07-15 18:12:54.077180] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.357 [2024-07-15 18:12:54.077195] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.357 [2024-07-15 18:12:54.077202] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.357 [2024-07-15 18:12:54.077208] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.357 [2024-07-15 18:12:54.077222] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.357 qpair failed and we were unable to recover it. 00:27:00.617 [2024-07-15 18:12:54.087153] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.617 [2024-07-15 18:12:54.087212] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.617 [2024-07-15 18:12:54.087232] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.617 [2024-07-15 18:12:54.087240] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.617 [2024-07-15 18:12:54.087246] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.617 [2024-07-15 18:12:54.087260] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.617 qpair failed and we were unable to recover it. 00:27:00.617 [2024-07-15 18:12:54.097207] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.617 [2024-07-15 18:12:54.097336] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.617 [2024-07-15 18:12:54.097353] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.617 [2024-07-15 18:12:54.097360] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.617 [2024-07-15 18:12:54.097367] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.617 [2024-07-15 18:12:54.097383] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.617 qpair failed and we were unable to recover it. 00:27:00.617 [2024-07-15 18:12:54.107227] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.617 [2024-07-15 18:12:54.107333] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.617 [2024-07-15 18:12:54.107349] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.617 [2024-07-15 18:12:54.107357] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.617 [2024-07-15 18:12:54.107363] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.617 [2024-07-15 18:12:54.107378] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.617 qpair failed and we were unable to recover it. 00:27:00.617 [2024-07-15 18:12:54.117252] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.617 [2024-07-15 18:12:54.117357] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.617 [2024-07-15 18:12:54.117372] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.617 [2024-07-15 18:12:54.117383] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.617 [2024-07-15 18:12:54.117390] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.617 [2024-07-15 18:12:54.117405] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.617 qpair failed and we were unable to recover it. 00:27:00.617 [2024-07-15 18:12:54.127191] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.617 [2024-07-15 18:12:54.127253] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.617 [2024-07-15 18:12:54.127268] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.617 [2024-07-15 18:12:54.127276] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.617 [2024-07-15 18:12:54.127282] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.617 [2024-07-15 18:12:54.127297] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.617 qpair failed and we were unable to recover it. 00:27:00.617 [2024-07-15 18:12:54.137276] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.617 [2024-07-15 18:12:54.137335] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.617 [2024-07-15 18:12:54.137351] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.617 [2024-07-15 18:12:54.137359] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.617 [2024-07-15 18:12:54.137365] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.617 [2024-07-15 18:12:54.137379] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.617 qpair failed and we were unable to recover it. 00:27:00.617 [2024-07-15 18:12:54.147331] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.617 [2024-07-15 18:12:54.147397] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.617 [2024-07-15 18:12:54.147412] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.617 [2024-07-15 18:12:54.147420] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.617 [2024-07-15 18:12:54.147426] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.617 [2024-07-15 18:12:54.147441] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.617 qpair failed and we were unable to recover it. 00:27:00.617 [2024-07-15 18:12:54.157286] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.617 [2024-07-15 18:12:54.157353] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.617 [2024-07-15 18:12:54.157369] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.617 [2024-07-15 18:12:54.157377] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.617 [2024-07-15 18:12:54.157384] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.617 [2024-07-15 18:12:54.157399] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.617 qpair failed and we were unable to recover it. 00:27:00.617 [2024-07-15 18:12:54.167399] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.617 [2024-07-15 18:12:54.167480] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.617 [2024-07-15 18:12:54.167496] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.617 [2024-07-15 18:12:54.167503] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.617 [2024-07-15 18:12:54.167510] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.617 [2024-07-15 18:12:54.167524] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.617 qpair failed and we were unable to recover it. 00:27:00.617 [2024-07-15 18:12:54.177401] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.617 [2024-07-15 18:12:54.177460] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.617 [2024-07-15 18:12:54.177476] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.617 [2024-07-15 18:12:54.177484] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.617 [2024-07-15 18:12:54.177491] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.617 [2024-07-15 18:12:54.177506] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.617 qpair failed and we were unable to recover it. 00:27:00.617 [2024-07-15 18:12:54.187438] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.617 [2024-07-15 18:12:54.187503] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.618 [2024-07-15 18:12:54.187519] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.618 [2024-07-15 18:12:54.187527] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.618 [2024-07-15 18:12:54.187533] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.618 [2024-07-15 18:12:54.187547] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.618 qpair failed and we were unable to recover it. 00:27:00.618 [2024-07-15 18:12:54.197423] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.618 [2024-07-15 18:12:54.197489] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.618 [2024-07-15 18:12:54.197505] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.618 [2024-07-15 18:12:54.197512] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.618 [2024-07-15 18:12:54.197518] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.618 [2024-07-15 18:12:54.197532] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.618 qpair failed and we were unable to recover it. 00:27:00.618 [2024-07-15 18:12:54.207531] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.618 [2024-07-15 18:12:54.207598] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.618 [2024-07-15 18:12:54.207618] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.618 [2024-07-15 18:12:54.207625] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.618 [2024-07-15 18:12:54.207632] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.618 [2024-07-15 18:12:54.207646] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.618 qpair failed and we were unable to recover it. 00:27:00.618 [2024-07-15 18:12:54.217510] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.618 [2024-07-15 18:12:54.217583] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.618 [2024-07-15 18:12:54.217599] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.618 [2024-07-15 18:12:54.217606] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.618 [2024-07-15 18:12:54.217613] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.618 [2024-07-15 18:12:54.217626] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.618 qpair failed and we were unable to recover it. 00:27:00.618 [2024-07-15 18:12:54.227573] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.618 [2024-07-15 18:12:54.227649] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.618 [2024-07-15 18:12:54.227665] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.618 [2024-07-15 18:12:54.227672] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.618 [2024-07-15 18:12:54.227679] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.618 [2024-07-15 18:12:54.227693] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.618 qpair failed and we were unable to recover it. 00:27:00.618 [2024-07-15 18:12:54.237603] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.618 [2024-07-15 18:12:54.237698] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.618 [2024-07-15 18:12:54.237713] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.618 [2024-07-15 18:12:54.237720] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.618 [2024-07-15 18:12:54.237726] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.618 [2024-07-15 18:12:54.237741] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.618 qpair failed and we were unable to recover it. 00:27:00.618 [2024-07-15 18:12:54.247545] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.618 [2024-07-15 18:12:54.247613] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.618 [2024-07-15 18:12:54.247630] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.618 [2024-07-15 18:12:54.247637] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.618 [2024-07-15 18:12:54.247643] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.618 [2024-07-15 18:12:54.247657] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.618 qpair failed and we were unable to recover it. 00:27:00.618 [2024-07-15 18:12:54.257657] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.618 [2024-07-15 18:12:54.257722] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.618 [2024-07-15 18:12:54.257739] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.618 [2024-07-15 18:12:54.257746] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.618 [2024-07-15 18:12:54.257752] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.618 [2024-07-15 18:12:54.257768] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.618 qpair failed and we were unable to recover it. 00:27:00.618 [2024-07-15 18:12:54.267670] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.618 [2024-07-15 18:12:54.267735] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.618 [2024-07-15 18:12:54.267751] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.618 [2024-07-15 18:12:54.267759] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.618 [2024-07-15 18:12:54.267765] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.618 [2024-07-15 18:12:54.267779] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.618 qpair failed and we were unable to recover it. 00:27:00.618 [2024-07-15 18:12:54.277686] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.618 [2024-07-15 18:12:54.277750] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.618 [2024-07-15 18:12:54.277765] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.618 [2024-07-15 18:12:54.277772] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.618 [2024-07-15 18:12:54.277778] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.618 [2024-07-15 18:12:54.277792] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.618 qpair failed and we were unable to recover it. 00:27:00.618 [2024-07-15 18:12:54.287718] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.618 [2024-07-15 18:12:54.287782] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.618 [2024-07-15 18:12:54.287799] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.618 [2024-07-15 18:12:54.287807] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.618 [2024-07-15 18:12:54.287813] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.618 [2024-07-15 18:12:54.287829] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.618 qpair failed and we were unable to recover it. 00:27:00.618 [2024-07-15 18:12:54.297755] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.618 [2024-07-15 18:12:54.297822] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.618 [2024-07-15 18:12:54.297843] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.618 [2024-07-15 18:12:54.297850] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.618 [2024-07-15 18:12:54.297856] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.618 [2024-07-15 18:12:54.297872] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.618 qpair failed and we were unable to recover it. 00:27:00.618 [2024-07-15 18:12:54.307783] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.618 [2024-07-15 18:12:54.307849] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.618 [2024-07-15 18:12:54.307867] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.618 [2024-07-15 18:12:54.307875] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.618 [2024-07-15 18:12:54.307881] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.618 [2024-07-15 18:12:54.307896] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.618 qpair failed and we were unable to recover it. 00:27:00.618 [2024-07-15 18:12:54.317845] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.618 [2024-07-15 18:12:54.317911] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.618 [2024-07-15 18:12:54.317928] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.618 [2024-07-15 18:12:54.317935] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.618 [2024-07-15 18:12:54.317942] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.618 [2024-07-15 18:12:54.317956] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.618 qpair failed and we were unable to recover it. 00:27:00.618 [2024-07-15 18:12:54.327783] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.618 [2024-07-15 18:12:54.327846] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.618 [2024-07-15 18:12:54.327862] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.618 [2024-07-15 18:12:54.327869] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.618 [2024-07-15 18:12:54.327875] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.619 [2024-07-15 18:12:54.327889] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.619 qpair failed and we were unable to recover it. 00:27:00.619 [2024-07-15 18:12:54.337860] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.619 [2024-07-15 18:12:54.337927] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.619 [2024-07-15 18:12:54.337944] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.619 [2024-07-15 18:12:54.337951] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.619 [2024-07-15 18:12:54.337957] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.619 [2024-07-15 18:12:54.337974] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.619 qpair failed and we were unable to recover it. 00:27:00.878 [2024-07-15 18:12:54.347866] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.878 [2024-07-15 18:12:54.347929] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.878 [2024-07-15 18:12:54.347945] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.878 [2024-07-15 18:12:54.347952] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.878 [2024-07-15 18:12:54.347958] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.878 [2024-07-15 18:12:54.347973] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.878 qpair failed and we were unable to recover it. 00:27:00.878 [2024-07-15 18:12:54.357922] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.878 [2024-07-15 18:12:54.357986] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.878 [2024-07-15 18:12:54.358003] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.878 [2024-07-15 18:12:54.358010] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.878 [2024-07-15 18:12:54.358017] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.878 [2024-07-15 18:12:54.358032] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.878 qpair failed and we were unable to recover it. 00:27:00.878 [2024-07-15 18:12:54.368002] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.878 [2024-07-15 18:12:54.368094] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.879 [2024-07-15 18:12:54.368110] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.879 [2024-07-15 18:12:54.368117] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.879 [2024-07-15 18:12:54.368124] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.879 [2024-07-15 18:12:54.368139] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.879 qpair failed and we were unable to recover it. 00:27:00.879 [2024-07-15 18:12:54.377966] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.879 [2024-07-15 18:12:54.378030] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.879 [2024-07-15 18:12:54.378052] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.879 [2024-07-15 18:12:54.378060] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.879 [2024-07-15 18:12:54.378066] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.879 [2024-07-15 18:12:54.378081] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.879 qpair failed and we were unable to recover it. 00:27:00.879 [2024-07-15 18:12:54.388063] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.879 [2024-07-15 18:12:54.388127] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.879 [2024-07-15 18:12:54.388145] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.879 [2024-07-15 18:12:54.388152] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.879 [2024-07-15 18:12:54.388158] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.879 [2024-07-15 18:12:54.388172] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.879 qpair failed and we were unable to recover it. 00:27:00.879 [2024-07-15 18:12:54.398038] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.879 [2024-07-15 18:12:54.398126] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.879 [2024-07-15 18:12:54.398143] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.879 [2024-07-15 18:12:54.398152] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.879 [2024-07-15 18:12:54.398159] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.879 [2024-07-15 18:12:54.398174] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.879 qpair failed and we were unable to recover it. 00:27:00.879 [2024-07-15 18:12:54.407998] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.879 [2024-07-15 18:12:54.408063] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.879 [2024-07-15 18:12:54.408078] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.879 [2024-07-15 18:12:54.408085] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.879 [2024-07-15 18:12:54.408091] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.879 [2024-07-15 18:12:54.408105] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.879 qpair failed and we were unable to recover it. 00:27:00.879 [2024-07-15 18:12:54.418062] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.879 [2024-07-15 18:12:54.418134] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.879 [2024-07-15 18:12:54.418150] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.879 [2024-07-15 18:12:54.418157] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.879 [2024-07-15 18:12:54.418162] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.879 [2024-07-15 18:12:54.418176] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.879 qpair failed and we were unable to recover it. 00:27:00.879 [2024-07-15 18:12:54.428095] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.879 [2024-07-15 18:12:54.428161] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.879 [2024-07-15 18:12:54.428176] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.879 [2024-07-15 18:12:54.428183] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.879 [2024-07-15 18:12:54.428189] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.879 [2024-07-15 18:12:54.428206] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.879 qpair failed and we were unable to recover it. 00:27:00.879 [2024-07-15 18:12:54.438102] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.879 [2024-07-15 18:12:54.438171] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.879 [2024-07-15 18:12:54.438186] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.879 [2024-07-15 18:12:54.438193] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.879 [2024-07-15 18:12:54.438199] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.879 [2024-07-15 18:12:54.438212] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.879 qpair failed and we were unable to recover it. 00:27:00.879 [2024-07-15 18:12:54.448179] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.879 [2024-07-15 18:12:54.448248] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.879 [2024-07-15 18:12:54.448264] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.879 [2024-07-15 18:12:54.448271] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.879 [2024-07-15 18:12:54.448277] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.879 [2024-07-15 18:12:54.448291] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.879 qpair failed and we were unable to recover it. 00:27:00.879 [2024-07-15 18:12:54.458205] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.879 [2024-07-15 18:12:54.458271] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.879 [2024-07-15 18:12:54.458286] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.879 [2024-07-15 18:12:54.458293] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.879 [2024-07-15 18:12:54.458298] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.879 [2024-07-15 18:12:54.458312] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.879 qpair failed and we were unable to recover it. 00:27:00.879 [2024-07-15 18:12:54.468220] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.879 [2024-07-15 18:12:54.468296] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.879 [2024-07-15 18:12:54.468311] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.879 [2024-07-15 18:12:54.468318] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.879 [2024-07-15 18:12:54.468324] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.879 [2024-07-15 18:12:54.468338] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.879 qpair failed and we were unable to recover it. 00:27:00.879 [2024-07-15 18:12:54.478284] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.879 [2024-07-15 18:12:54.478347] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.879 [2024-07-15 18:12:54.478367] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.879 [2024-07-15 18:12:54.478374] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.879 [2024-07-15 18:12:54.478380] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.879 [2024-07-15 18:12:54.478394] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.879 qpair failed and we were unable to recover it. 00:27:00.879 [2024-07-15 18:12:54.488241] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.879 [2024-07-15 18:12:54.488301] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.879 [2024-07-15 18:12:54.488316] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.879 [2024-07-15 18:12:54.488323] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.879 [2024-07-15 18:12:54.488329] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.879 [2024-07-15 18:12:54.488343] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.879 qpair failed and we were unable to recover it. 00:27:00.879 [2024-07-15 18:12:54.498280] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.879 [2024-07-15 18:12:54.498343] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.879 [2024-07-15 18:12:54.498360] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.879 [2024-07-15 18:12:54.498366] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.879 [2024-07-15 18:12:54.498372] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.879 [2024-07-15 18:12:54.498387] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.879 qpair failed and we were unable to recover it. 00:27:00.879 [2024-07-15 18:12:54.508383] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.879 [2024-07-15 18:12:54.508448] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.879 [2024-07-15 18:12:54.508464] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.879 [2024-07-15 18:12:54.508470] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.879 [2024-07-15 18:12:54.508476] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.879 [2024-07-15 18:12:54.508489] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.879 qpair failed and we were unable to recover it. 00:27:00.879 [2024-07-15 18:12:54.518334] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.879 [2024-07-15 18:12:54.518400] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.879 [2024-07-15 18:12:54.518415] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.879 [2024-07-15 18:12:54.518422] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.879 [2024-07-15 18:12:54.518431] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.879 [2024-07-15 18:12:54.518444] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.879 qpair failed and we were unable to recover it. 00:27:00.879 [2024-07-15 18:12:54.528432] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.879 [2024-07-15 18:12:54.528497] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.879 [2024-07-15 18:12:54.528513] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.879 [2024-07-15 18:12:54.528519] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.879 [2024-07-15 18:12:54.528525] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.879 [2024-07-15 18:12:54.528538] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.879 qpair failed and we were unable to recover it. 00:27:00.879 [2024-07-15 18:12:54.538460] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.879 [2024-07-15 18:12:54.538523] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.879 [2024-07-15 18:12:54.538538] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.879 [2024-07-15 18:12:54.538545] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.879 [2024-07-15 18:12:54.538551] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.879 [2024-07-15 18:12:54.538564] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.879 qpair failed and we were unable to recover it. 00:27:00.879 [2024-07-15 18:12:54.548515] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.879 [2024-07-15 18:12:54.548583] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.879 [2024-07-15 18:12:54.548598] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.879 [2024-07-15 18:12:54.548605] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.879 [2024-07-15 18:12:54.548611] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.879 [2024-07-15 18:12:54.548625] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.879 qpair failed and we were unable to recover it. 00:27:00.879 [2024-07-15 18:12:54.558526] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.880 [2024-07-15 18:12:54.558588] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.880 [2024-07-15 18:12:54.558603] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.880 [2024-07-15 18:12:54.558610] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.880 [2024-07-15 18:12:54.558615] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.880 [2024-07-15 18:12:54.558629] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.880 qpair failed and we were unable to recover it. 00:27:00.880 [2024-07-15 18:12:54.568491] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.880 [2024-07-15 18:12:54.568555] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.880 [2024-07-15 18:12:54.568570] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.880 [2024-07-15 18:12:54.568577] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.880 [2024-07-15 18:12:54.568583] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.880 [2024-07-15 18:12:54.568597] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.880 qpair failed and we were unable to recover it. 00:27:00.880 [2024-07-15 18:12:54.578600] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.880 [2024-07-15 18:12:54.578670] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.880 [2024-07-15 18:12:54.578687] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.880 [2024-07-15 18:12:54.578694] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.880 [2024-07-15 18:12:54.578700] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.880 [2024-07-15 18:12:54.578715] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.880 qpair failed and we were unable to recover it. 00:27:00.880 [2024-07-15 18:12:54.588592] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.880 [2024-07-15 18:12:54.588653] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.880 [2024-07-15 18:12:54.588669] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.880 [2024-07-15 18:12:54.588676] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.880 [2024-07-15 18:12:54.588681] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.880 [2024-07-15 18:12:54.588696] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.880 qpair failed and we were unable to recover it. 00:27:00.880 [2024-07-15 18:12:54.598608] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:00.880 [2024-07-15 18:12:54.598669] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:00.880 [2024-07-15 18:12:54.598685] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:00.880 [2024-07-15 18:12:54.598691] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:00.880 [2024-07-15 18:12:54.598697] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:00.880 [2024-07-15 18:12:54.598711] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:00.880 qpair failed and we were unable to recover it. 00:27:01.139 [2024-07-15 18:12:54.608661] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.139 [2024-07-15 18:12:54.608723] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.139 [2024-07-15 18:12:54.608738] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.139 [2024-07-15 18:12:54.608745] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.139 [2024-07-15 18:12:54.608756] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.139 [2024-07-15 18:12:54.608770] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.139 qpair failed and we were unable to recover it. 00:27:01.139 [2024-07-15 18:12:54.618688] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.139 [2024-07-15 18:12:54.618751] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.139 [2024-07-15 18:12:54.618766] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.139 [2024-07-15 18:12:54.618772] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.139 [2024-07-15 18:12:54.618778] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.139 [2024-07-15 18:12:54.618791] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.139 qpair failed and we were unable to recover it. 00:27:01.139 [2024-07-15 18:12:54.628751] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.139 [2024-07-15 18:12:54.628814] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.139 [2024-07-15 18:12:54.628828] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.139 [2024-07-15 18:12:54.628834] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.139 [2024-07-15 18:12:54.628840] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.139 [2024-07-15 18:12:54.628854] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.139 qpair failed and we were unable to recover it. 00:27:01.140 [2024-07-15 18:12:54.638773] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.140 [2024-07-15 18:12:54.638837] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.140 [2024-07-15 18:12:54.638852] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.140 [2024-07-15 18:12:54.638858] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.140 [2024-07-15 18:12:54.638864] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.140 [2024-07-15 18:12:54.638877] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.140 qpair failed and we were unable to recover it. 00:27:01.140 [2024-07-15 18:12:54.648776] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.140 [2024-07-15 18:12:54.648838] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.140 [2024-07-15 18:12:54.648853] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.140 [2024-07-15 18:12:54.648859] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.140 [2024-07-15 18:12:54.648865] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.140 [2024-07-15 18:12:54.648878] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.140 qpair failed and we were unable to recover it. 00:27:01.140 [2024-07-15 18:12:54.658806] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.140 [2024-07-15 18:12:54.658868] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.140 [2024-07-15 18:12:54.658884] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.140 [2024-07-15 18:12:54.658890] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.140 [2024-07-15 18:12:54.658896] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.140 [2024-07-15 18:12:54.658909] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.140 qpair failed and we were unable to recover it. 00:27:01.140 [2024-07-15 18:12:54.668841] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.140 [2024-07-15 18:12:54.668905] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.140 [2024-07-15 18:12:54.668921] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.140 [2024-07-15 18:12:54.668928] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.140 [2024-07-15 18:12:54.668933] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.140 [2024-07-15 18:12:54.668947] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.140 qpair failed and we were unable to recover it. 00:27:01.140 [2024-07-15 18:12:54.678876] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.140 [2024-07-15 18:12:54.678941] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.140 [2024-07-15 18:12:54.678957] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.140 [2024-07-15 18:12:54.678964] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.140 [2024-07-15 18:12:54.678970] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.140 [2024-07-15 18:12:54.678984] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.140 qpair failed and we were unable to recover it. 00:27:01.140 [2024-07-15 18:12:54.688875] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.140 [2024-07-15 18:12:54.688931] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.140 [2024-07-15 18:12:54.688946] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.140 [2024-07-15 18:12:54.688953] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.140 [2024-07-15 18:12:54.688959] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.140 [2024-07-15 18:12:54.688973] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.140 qpair failed and we were unable to recover it. 00:27:01.140 [2024-07-15 18:12:54.698880] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.140 [2024-07-15 18:12:54.698947] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.140 [2024-07-15 18:12:54.698962] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.140 [2024-07-15 18:12:54.698969] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.140 [2024-07-15 18:12:54.698978] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.140 [2024-07-15 18:12:54.698991] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.140 qpair failed and we were unable to recover it. 00:27:01.140 [2024-07-15 18:12:54.708940] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.140 [2024-07-15 18:12:54.709003] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.140 [2024-07-15 18:12:54.709019] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.140 [2024-07-15 18:12:54.709026] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.140 [2024-07-15 18:12:54.709031] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.140 [2024-07-15 18:12:54.709045] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.140 qpair failed and we were unable to recover it. 00:27:01.140 [2024-07-15 18:12:54.718964] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.140 [2024-07-15 18:12:54.719025] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.140 [2024-07-15 18:12:54.719040] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.140 [2024-07-15 18:12:54.719046] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.140 [2024-07-15 18:12:54.719052] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.140 [2024-07-15 18:12:54.719066] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.140 qpair failed and we were unable to recover it. 00:27:01.140 [2024-07-15 18:12:54.729012] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.140 [2024-07-15 18:12:54.729074] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.140 [2024-07-15 18:12:54.729089] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.140 [2024-07-15 18:12:54.729096] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.140 [2024-07-15 18:12:54.729102] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.140 [2024-07-15 18:12:54.729115] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.140 qpair failed and we were unable to recover it. 00:27:01.140 [2024-07-15 18:12:54.739036] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.140 [2024-07-15 18:12:54.739096] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.140 [2024-07-15 18:12:54.739113] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.140 [2024-07-15 18:12:54.739120] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.140 [2024-07-15 18:12:54.739126] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.140 [2024-07-15 18:12:54.739140] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.140 qpair failed and we were unable to recover it. 00:27:01.140 [2024-07-15 18:12:54.749095] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.140 [2024-07-15 18:12:54.749158] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.140 [2024-07-15 18:12:54.749173] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.140 [2024-07-15 18:12:54.749181] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.140 [2024-07-15 18:12:54.749186] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.140 [2024-07-15 18:12:54.749200] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.140 qpair failed and we were unable to recover it. 00:27:01.140 [2024-07-15 18:12:54.759083] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.140 [2024-07-15 18:12:54.759145] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.140 [2024-07-15 18:12:54.759162] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.140 [2024-07-15 18:12:54.759169] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.140 [2024-07-15 18:12:54.759175] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.140 [2024-07-15 18:12:54.759189] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.140 qpair failed and we were unable to recover it. 00:27:01.140 [2024-07-15 18:12:54.769126] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.140 [2024-07-15 18:12:54.769193] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.140 [2024-07-15 18:12:54.769208] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.140 [2024-07-15 18:12:54.769215] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.140 [2024-07-15 18:12:54.769221] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.140 [2024-07-15 18:12:54.769240] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.140 qpair failed and we were unable to recover it. 00:27:01.140 [2024-07-15 18:12:54.779153] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.140 [2024-07-15 18:12:54.779219] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.140 [2024-07-15 18:12:54.779238] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.141 [2024-07-15 18:12:54.779245] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.141 [2024-07-15 18:12:54.779251] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.141 [2024-07-15 18:12:54.779265] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.141 qpair failed and we were unable to recover it. 00:27:01.141 [2024-07-15 18:12:54.789163] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.141 [2024-07-15 18:12:54.789232] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.141 [2024-07-15 18:12:54.789247] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.141 [2024-07-15 18:12:54.789258] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.141 [2024-07-15 18:12:54.789263] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.141 [2024-07-15 18:12:54.789277] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.141 qpair failed and we were unable to recover it. 00:27:01.141 [2024-07-15 18:12:54.799227] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.141 [2024-07-15 18:12:54.799300] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.141 [2024-07-15 18:12:54.799315] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.141 [2024-07-15 18:12:54.799322] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.141 [2024-07-15 18:12:54.799328] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.141 [2024-07-15 18:12:54.799343] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.141 qpair failed and we were unable to recover it. 00:27:01.141 [2024-07-15 18:12:54.809226] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.141 [2024-07-15 18:12:54.809290] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.141 [2024-07-15 18:12:54.809305] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.141 [2024-07-15 18:12:54.809312] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.141 [2024-07-15 18:12:54.809317] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.141 [2024-07-15 18:12:54.809331] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.141 qpair failed and we were unable to recover it. 00:27:01.141 [2024-07-15 18:12:54.819278] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.141 [2024-07-15 18:12:54.819341] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.141 [2024-07-15 18:12:54.819357] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.141 [2024-07-15 18:12:54.819363] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.141 [2024-07-15 18:12:54.819369] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.141 [2024-07-15 18:12:54.819383] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.141 qpair failed and we were unable to recover it. 00:27:01.141 [2024-07-15 18:12:54.829343] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.141 [2024-07-15 18:12:54.829405] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.141 [2024-07-15 18:12:54.829420] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.141 [2024-07-15 18:12:54.829427] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.141 [2024-07-15 18:12:54.829433] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.141 [2024-07-15 18:12:54.829446] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.141 qpair failed and we were unable to recover it. 00:27:01.141 [2024-07-15 18:12:54.839336] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.141 [2024-07-15 18:12:54.839422] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.141 [2024-07-15 18:12:54.839438] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.141 [2024-07-15 18:12:54.839445] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.141 [2024-07-15 18:12:54.839451] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.141 [2024-07-15 18:12:54.839465] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.141 qpair failed and we were unable to recover it. 00:27:01.141 [2024-07-15 18:12:54.849350] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.141 [2024-07-15 18:12:54.849410] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.141 [2024-07-15 18:12:54.849425] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.141 [2024-07-15 18:12:54.849432] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.141 [2024-07-15 18:12:54.849438] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.141 [2024-07-15 18:12:54.849452] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.141 qpair failed and we were unable to recover it. 00:27:01.141 [2024-07-15 18:12:54.859414] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.141 [2024-07-15 18:12:54.859521] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.141 [2024-07-15 18:12:54.859537] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.141 [2024-07-15 18:12:54.859544] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.141 [2024-07-15 18:12:54.859551] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.141 [2024-07-15 18:12:54.859565] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.141 qpair failed and we were unable to recover it. 00:27:01.401 [2024-07-15 18:12:54.869406] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.401 [2024-07-15 18:12:54.869468] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.401 [2024-07-15 18:12:54.869483] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.401 [2024-07-15 18:12:54.869490] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.401 [2024-07-15 18:12:54.869496] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.401 [2024-07-15 18:12:54.869509] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.401 qpair failed and we were unable to recover it. 00:27:01.401 [2024-07-15 18:12:54.879443] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.401 [2024-07-15 18:12:54.879508] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.401 [2024-07-15 18:12:54.879523] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.401 [2024-07-15 18:12:54.879533] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.401 [2024-07-15 18:12:54.879540] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.401 [2024-07-15 18:12:54.879553] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.401 qpair failed and we were unable to recover it. 00:27:01.401 [2024-07-15 18:12:54.889462] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.401 [2024-07-15 18:12:54.889526] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.401 [2024-07-15 18:12:54.889540] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.401 [2024-07-15 18:12:54.889547] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.401 [2024-07-15 18:12:54.889553] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.401 [2024-07-15 18:12:54.889566] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.401 qpair failed and we were unable to recover it. 00:27:01.401 [2024-07-15 18:12:54.899509] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.401 [2024-07-15 18:12:54.899568] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.401 [2024-07-15 18:12:54.899584] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.401 [2024-07-15 18:12:54.899591] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.401 [2024-07-15 18:12:54.899596] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.401 [2024-07-15 18:12:54.899610] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.401 qpair failed and we were unable to recover it. 00:27:01.401 [2024-07-15 18:12:54.909512] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.401 [2024-07-15 18:12:54.909574] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.401 [2024-07-15 18:12:54.909589] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.401 [2024-07-15 18:12:54.909596] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.401 [2024-07-15 18:12:54.909602] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.401 [2024-07-15 18:12:54.909615] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.401 qpair failed and we were unable to recover it. 00:27:01.401 [2024-07-15 18:12:54.919580] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.401 [2024-07-15 18:12:54.919688] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.401 [2024-07-15 18:12:54.919704] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.401 [2024-07-15 18:12:54.919711] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.401 [2024-07-15 18:12:54.919717] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.401 [2024-07-15 18:12:54.919731] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.401 qpair failed and we were unable to recover it. 00:27:01.401 [2024-07-15 18:12:54.929588] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.401 [2024-07-15 18:12:54.929655] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.401 [2024-07-15 18:12:54.929671] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.401 [2024-07-15 18:12:54.929677] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.401 [2024-07-15 18:12:54.929683] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.401 [2024-07-15 18:12:54.929697] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.401 qpair failed and we were unable to recover it. 00:27:01.401 [2024-07-15 18:12:54.939589] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.401 [2024-07-15 18:12:54.939646] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.401 [2024-07-15 18:12:54.939661] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.401 [2024-07-15 18:12:54.939667] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.401 [2024-07-15 18:12:54.939673] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.401 [2024-07-15 18:12:54.939687] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.401 qpair failed and we were unable to recover it. 00:27:01.401 [2024-07-15 18:12:54.949654] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.401 [2024-07-15 18:12:54.949714] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.401 [2024-07-15 18:12:54.949729] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.401 [2024-07-15 18:12:54.949735] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.401 [2024-07-15 18:12:54.949741] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.401 [2024-07-15 18:12:54.949755] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.401 qpair failed and we were unable to recover it. 00:27:01.401 [2024-07-15 18:12:54.959638] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.402 [2024-07-15 18:12:54.959704] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.402 [2024-07-15 18:12:54.959719] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.402 [2024-07-15 18:12:54.959726] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.402 [2024-07-15 18:12:54.959731] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.402 [2024-07-15 18:12:54.959746] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.402 qpair failed and we were unable to recover it. 00:27:01.402 [2024-07-15 18:12:54.969626] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.402 [2024-07-15 18:12:54.969689] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.402 [2024-07-15 18:12:54.969704] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.402 [2024-07-15 18:12:54.969714] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.402 [2024-07-15 18:12:54.969719] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.402 [2024-07-15 18:12:54.969733] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.402 qpair failed and we were unable to recover it. 00:27:01.402 [2024-07-15 18:12:54.979718] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.402 [2024-07-15 18:12:54.979783] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.402 [2024-07-15 18:12:54.979799] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.402 [2024-07-15 18:12:54.979805] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.402 [2024-07-15 18:12:54.979812] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.402 [2024-07-15 18:12:54.979826] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.402 qpair failed and we were unable to recover it. 00:27:01.402 [2024-07-15 18:12:54.989797] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.402 [2024-07-15 18:12:54.989859] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.402 [2024-07-15 18:12:54.989874] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.402 [2024-07-15 18:12:54.989881] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.402 [2024-07-15 18:12:54.989886] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.402 [2024-07-15 18:12:54.989900] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.402 qpair failed and we were unable to recover it. 00:27:01.402 [2024-07-15 18:12:54.999775] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.402 [2024-07-15 18:12:54.999838] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.402 [2024-07-15 18:12:54.999853] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.402 [2024-07-15 18:12:54.999859] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.402 [2024-07-15 18:12:54.999865] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.402 [2024-07-15 18:12:54.999878] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.402 qpair failed and we were unable to recover it. 00:27:01.402 [2024-07-15 18:12:55.009827] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.402 [2024-07-15 18:12:55.009923] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.402 [2024-07-15 18:12:55.009939] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.402 [2024-07-15 18:12:55.009946] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.402 [2024-07-15 18:12:55.009952] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.402 [2024-07-15 18:12:55.009967] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.402 qpair failed and we were unable to recover it. 00:27:01.402 [2024-07-15 18:12:55.019853] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.402 [2024-07-15 18:12:55.019953] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.402 [2024-07-15 18:12:55.019968] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.402 [2024-07-15 18:12:55.019975] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.402 [2024-07-15 18:12:55.019981] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.402 [2024-07-15 18:12:55.019996] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.402 qpair failed and we were unable to recover it. 00:27:01.402 [2024-07-15 18:12:55.029916] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.402 [2024-07-15 18:12:55.029982] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.402 [2024-07-15 18:12:55.029998] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.402 [2024-07-15 18:12:55.030004] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.402 [2024-07-15 18:12:55.030010] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.402 [2024-07-15 18:12:55.030025] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.402 qpair failed and we were unable to recover it. 00:27:01.402 [2024-07-15 18:12:55.039912] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.402 [2024-07-15 18:12:55.039973] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.402 [2024-07-15 18:12:55.039988] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.402 [2024-07-15 18:12:55.039995] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.402 [2024-07-15 18:12:55.040001] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.402 [2024-07-15 18:12:55.040014] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.402 qpair failed and we were unable to recover it. 00:27:01.402 [2024-07-15 18:12:55.049916] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.402 [2024-07-15 18:12:55.049977] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.402 [2024-07-15 18:12:55.049992] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.402 [2024-07-15 18:12:55.049999] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.402 [2024-07-15 18:12:55.050005] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.402 [2024-07-15 18:12:55.050018] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.402 qpair failed and we were unable to recover it. 00:27:01.402 [2024-07-15 18:12:55.059994] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.402 [2024-07-15 18:12:55.060059] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.402 [2024-07-15 18:12:55.060078] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.402 [2024-07-15 18:12:55.060085] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.402 [2024-07-15 18:12:55.060091] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.402 [2024-07-15 18:12:55.060105] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.402 qpair failed and we were unable to recover it. 00:27:01.402 [2024-07-15 18:12:55.069995] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.402 [2024-07-15 18:12:55.070104] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.402 [2024-07-15 18:12:55.070124] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.402 [2024-07-15 18:12:55.070132] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.402 [2024-07-15 18:12:55.070138] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.402 [2024-07-15 18:12:55.070153] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.402 qpair failed and we were unable to recover it. 00:27:01.402 [2024-07-15 18:12:55.079949] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.402 [2024-07-15 18:12:55.080013] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.402 [2024-07-15 18:12:55.080029] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.402 [2024-07-15 18:12:55.080037] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.402 [2024-07-15 18:12:55.080043] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.402 [2024-07-15 18:12:55.080056] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.402 qpair failed and we were unable to recover it. 00:27:01.402 [2024-07-15 18:12:55.090041] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.402 [2024-07-15 18:12:55.090104] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.402 [2024-07-15 18:12:55.090120] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.402 [2024-07-15 18:12:55.090126] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.403 [2024-07-15 18:12:55.090132] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.403 [2024-07-15 18:12:55.090146] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.403 qpair failed and we were unable to recover it. 00:27:01.403 [2024-07-15 18:12:55.100088] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.403 [2024-07-15 18:12:55.100151] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.403 [2024-07-15 18:12:55.100167] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.403 [2024-07-15 18:12:55.100174] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.403 [2024-07-15 18:12:55.100179] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.403 [2024-07-15 18:12:55.100197] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.403 qpair failed and we were unable to recover it. 00:27:01.403 [2024-07-15 18:12:55.110095] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.403 [2024-07-15 18:12:55.110156] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.403 [2024-07-15 18:12:55.110172] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.403 [2024-07-15 18:12:55.110179] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.403 [2024-07-15 18:12:55.110184] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.403 [2024-07-15 18:12:55.110198] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.403 qpair failed and we were unable to recover it. 00:27:01.403 [2024-07-15 18:12:55.120104] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.403 [2024-07-15 18:12:55.120165] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.403 [2024-07-15 18:12:55.120180] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.403 [2024-07-15 18:12:55.120187] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.403 [2024-07-15 18:12:55.120192] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.403 [2024-07-15 18:12:55.120206] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.403 qpair failed and we were unable to recover it. 00:27:01.663 [2024-07-15 18:12:55.130144] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.663 [2024-07-15 18:12:55.130201] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.663 [2024-07-15 18:12:55.130216] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.663 [2024-07-15 18:12:55.130223] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.663 [2024-07-15 18:12:55.130232] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.663 [2024-07-15 18:12:55.130246] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.663 qpair failed and we were unable to recover it. 00:27:01.663 [2024-07-15 18:12:55.140181] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.663 [2024-07-15 18:12:55.140243] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.663 [2024-07-15 18:12:55.140258] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.663 [2024-07-15 18:12:55.140265] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.663 [2024-07-15 18:12:55.140271] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.663 [2024-07-15 18:12:55.140285] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.663 qpair failed and we were unable to recover it. 00:27:01.663 [2024-07-15 18:12:55.150233] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.663 [2024-07-15 18:12:55.150300] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.663 [2024-07-15 18:12:55.150318] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.663 [2024-07-15 18:12:55.150325] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.663 [2024-07-15 18:12:55.150331] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.663 [2024-07-15 18:12:55.150344] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.663 qpair failed and we were unable to recover it. 00:27:01.663 [2024-07-15 18:12:55.160247] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.663 [2024-07-15 18:12:55.160308] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.663 [2024-07-15 18:12:55.160324] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.663 [2024-07-15 18:12:55.160331] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.663 [2024-07-15 18:12:55.160337] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.663 [2024-07-15 18:12:55.160351] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.663 qpair failed and we were unable to recover it. 00:27:01.663 [2024-07-15 18:12:55.170196] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.663 [2024-07-15 18:12:55.170262] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.663 [2024-07-15 18:12:55.170277] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.663 [2024-07-15 18:12:55.170284] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.663 [2024-07-15 18:12:55.170290] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.663 [2024-07-15 18:12:55.170304] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.663 qpair failed and we were unable to recover it. 00:27:01.663 [2024-07-15 18:12:55.180281] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.663 [2024-07-15 18:12:55.180342] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.663 [2024-07-15 18:12:55.180357] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.663 [2024-07-15 18:12:55.180364] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.663 [2024-07-15 18:12:55.180370] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.663 [2024-07-15 18:12:55.180384] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.663 qpair failed and we were unable to recover it. 00:27:01.663 [2024-07-15 18:12:55.190340] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.663 [2024-07-15 18:12:55.190403] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.663 [2024-07-15 18:12:55.190418] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.663 [2024-07-15 18:12:55.190425] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.663 [2024-07-15 18:12:55.190431] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.663 [2024-07-15 18:12:55.190448] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.663 qpair failed and we were unable to recover it. 00:27:01.663 [2024-07-15 18:12:55.200388] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.663 [2024-07-15 18:12:55.200497] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.663 [2024-07-15 18:12:55.200513] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.663 [2024-07-15 18:12:55.200520] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.663 [2024-07-15 18:12:55.200526] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.663 [2024-07-15 18:12:55.200540] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.663 qpair failed and we were unable to recover it. 00:27:01.664 [2024-07-15 18:12:55.210436] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.664 [2024-07-15 18:12:55.210499] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.664 [2024-07-15 18:12:55.210515] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.664 [2024-07-15 18:12:55.210521] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.664 [2024-07-15 18:12:55.210527] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.664 [2024-07-15 18:12:55.210540] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.664 qpair failed and we were unable to recover it. 00:27:01.664 [2024-07-15 18:12:55.220406] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.664 [2024-07-15 18:12:55.220471] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.664 [2024-07-15 18:12:55.220487] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.664 [2024-07-15 18:12:55.220494] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.664 [2024-07-15 18:12:55.220500] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.664 [2024-07-15 18:12:55.220514] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.664 qpair failed and we were unable to recover it. 00:27:01.664 [2024-07-15 18:12:55.230450] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.664 [2024-07-15 18:12:55.230514] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.664 [2024-07-15 18:12:55.230529] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.664 [2024-07-15 18:12:55.230536] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.664 [2024-07-15 18:12:55.230543] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.664 [2024-07-15 18:12:55.230556] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.664 qpair failed and we were unable to recover it. 00:27:01.664 [2024-07-15 18:12:55.240459] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.664 [2024-07-15 18:12:55.240522] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.664 [2024-07-15 18:12:55.240541] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.664 [2024-07-15 18:12:55.240549] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.664 [2024-07-15 18:12:55.240555] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.664 [2024-07-15 18:12:55.240569] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.664 qpair failed and we were unable to recover it. 00:27:01.664 [2024-07-15 18:12:55.250505] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.664 [2024-07-15 18:12:55.250565] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.664 [2024-07-15 18:12:55.250581] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.664 [2024-07-15 18:12:55.250587] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.664 [2024-07-15 18:12:55.250593] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.664 [2024-07-15 18:12:55.250607] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.664 qpair failed and we were unable to recover it. 00:27:01.664 [2024-07-15 18:12:55.260537] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.664 [2024-07-15 18:12:55.260593] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.664 [2024-07-15 18:12:55.260608] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.664 [2024-07-15 18:12:55.260615] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.664 [2024-07-15 18:12:55.260620] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.664 [2024-07-15 18:12:55.260633] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.664 qpair failed and we were unable to recover it. 00:27:01.664 [2024-07-15 18:12:55.270488] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.664 [2024-07-15 18:12:55.270559] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.664 [2024-07-15 18:12:55.270574] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.664 [2024-07-15 18:12:55.270581] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.664 [2024-07-15 18:12:55.270586] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.664 [2024-07-15 18:12:55.270600] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.664 qpair failed and we were unable to recover it. 00:27:01.664 [2024-07-15 18:12:55.280632] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.664 [2024-07-15 18:12:55.280699] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.664 [2024-07-15 18:12:55.280715] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.664 [2024-07-15 18:12:55.280722] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.664 [2024-07-15 18:12:55.280727] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.664 [2024-07-15 18:12:55.280926] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.664 qpair failed and we were unable to recover it. 00:27:01.664 [2024-07-15 18:12:55.290593] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.664 [2024-07-15 18:12:55.290658] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.664 [2024-07-15 18:12:55.290673] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.664 [2024-07-15 18:12:55.290680] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.664 [2024-07-15 18:12:55.290686] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.664 [2024-07-15 18:12:55.290699] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.664 qpair failed and we were unable to recover it. 00:27:01.664 [2024-07-15 18:12:55.300650] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.664 [2024-07-15 18:12:55.300712] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.664 [2024-07-15 18:12:55.300728] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.664 [2024-07-15 18:12:55.300735] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.664 [2024-07-15 18:12:55.300740] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.664 [2024-07-15 18:12:55.300754] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.664 qpair failed and we were unable to recover it. 00:27:01.664 [2024-07-15 18:12:55.310662] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.664 [2024-07-15 18:12:55.310723] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.664 [2024-07-15 18:12:55.310739] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.664 [2024-07-15 18:12:55.310745] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.664 [2024-07-15 18:12:55.310751] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.664 [2024-07-15 18:12:55.310765] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.664 qpair failed and we were unable to recover it. 00:27:01.664 [2024-07-15 18:12:55.320729] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.664 [2024-07-15 18:12:55.320836] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.664 [2024-07-15 18:12:55.320857] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.664 [2024-07-15 18:12:55.320864] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.664 [2024-07-15 18:12:55.320870] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.664 [2024-07-15 18:12:55.320884] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.664 qpair failed and we were unable to recover it. 00:27:01.664 [2024-07-15 18:12:55.330746] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.664 [2024-07-15 18:12:55.330813] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.664 [2024-07-15 18:12:55.330834] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.664 [2024-07-15 18:12:55.330841] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.664 [2024-07-15 18:12:55.330846] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.664 [2024-07-15 18:12:55.330860] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.664 qpair failed and we were unable to recover it. 00:27:01.664 [2024-07-15 18:12:55.340762] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.664 [2024-07-15 18:12:55.340823] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.664 [2024-07-15 18:12:55.340838] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.664 [2024-07-15 18:12:55.340845] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.664 [2024-07-15 18:12:55.340851] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.664 [2024-07-15 18:12:55.340864] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.664 qpair failed and we were unable to recover it. 00:27:01.664 [2024-07-15 18:12:55.350799] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.665 [2024-07-15 18:12:55.350868] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.665 [2024-07-15 18:12:55.350883] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.665 [2024-07-15 18:12:55.350890] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.665 [2024-07-15 18:12:55.350895] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.665 [2024-07-15 18:12:55.350909] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.665 qpair failed and we were unable to recover it. 00:27:01.665 [2024-07-15 18:12:55.360829] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.665 [2024-07-15 18:12:55.360889] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.665 [2024-07-15 18:12:55.360905] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.665 [2024-07-15 18:12:55.360911] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.665 [2024-07-15 18:12:55.360917] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.665 [2024-07-15 18:12:55.360930] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.665 qpair failed and we were unable to recover it. 00:27:01.665 [2024-07-15 18:12:55.370821] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.665 [2024-07-15 18:12:55.370891] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.665 [2024-07-15 18:12:55.370906] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.665 [2024-07-15 18:12:55.370913] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.665 [2024-07-15 18:12:55.370922] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.665 [2024-07-15 18:12:55.370936] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.665 qpair failed and we were unable to recover it. 00:27:01.665 [2024-07-15 18:12:55.380799] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.665 [2024-07-15 18:12:55.380860] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.665 [2024-07-15 18:12:55.380875] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.665 [2024-07-15 18:12:55.380882] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.665 [2024-07-15 18:12:55.380888] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.665 [2024-07-15 18:12:55.380901] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.665 qpair failed and we were unable to recover it. 00:27:01.925 [2024-07-15 18:12:55.390918] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.925 [2024-07-15 18:12:55.390982] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.925 [2024-07-15 18:12:55.390997] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.925 [2024-07-15 18:12:55.391004] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.925 [2024-07-15 18:12:55.391010] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.925 [2024-07-15 18:12:55.391023] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.925 qpair failed and we were unable to recover it. 00:27:01.925 [2024-07-15 18:12:55.400938] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.925 [2024-07-15 18:12:55.401000] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.925 [2024-07-15 18:12:55.401017] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.925 [2024-07-15 18:12:55.401024] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.925 [2024-07-15 18:12:55.401031] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.925 [2024-07-15 18:12:55.401045] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.925 qpair failed and we were unable to recover it. 00:27:01.925 [2024-07-15 18:12:55.410949] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.925 [2024-07-15 18:12:55.411009] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.925 [2024-07-15 18:12:55.411025] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.925 [2024-07-15 18:12:55.411031] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.925 [2024-07-15 18:12:55.411037] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.925 [2024-07-15 18:12:55.411051] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.925 qpair failed and we were unable to recover it. 00:27:01.925 [2024-07-15 18:12:55.420979] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.925 [2024-07-15 18:12:55.421039] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.925 [2024-07-15 18:12:55.421055] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.925 [2024-07-15 18:12:55.421061] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.925 [2024-07-15 18:12:55.421067] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.925 [2024-07-15 18:12:55.421081] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.925 qpair failed and we were unable to recover it. 00:27:01.925 [2024-07-15 18:12:55.431010] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.925 [2024-07-15 18:12:55.431073] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.925 [2024-07-15 18:12:55.431089] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.925 [2024-07-15 18:12:55.431095] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.925 [2024-07-15 18:12:55.431101] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.925 [2024-07-15 18:12:55.431114] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.925 qpair failed and we were unable to recover it. 00:27:01.925 [2024-07-15 18:12:55.441006] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.925 [2024-07-15 18:12:55.441074] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.925 [2024-07-15 18:12:55.441089] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.925 [2024-07-15 18:12:55.441096] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.925 [2024-07-15 18:12:55.441101] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.925 [2024-07-15 18:12:55.441115] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.925 qpair failed and we were unable to recover it. 00:27:01.925 [2024-07-15 18:12:55.451076] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.925 [2024-07-15 18:12:55.451156] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.925 [2024-07-15 18:12:55.451172] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.925 [2024-07-15 18:12:55.451180] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.925 [2024-07-15 18:12:55.451186] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.925 [2024-07-15 18:12:55.451200] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.925 qpair failed and we were unable to recover it. 00:27:01.925 [2024-07-15 18:12:55.461087] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.925 [2024-07-15 18:12:55.461193] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.925 [2024-07-15 18:12:55.461209] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.925 [2024-07-15 18:12:55.461216] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.925 [2024-07-15 18:12:55.461229] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.925 [2024-07-15 18:12:55.461244] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.925 qpair failed and we were unable to recover it. 00:27:01.925 [2024-07-15 18:12:55.471132] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.925 [2024-07-15 18:12:55.471195] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.925 [2024-07-15 18:12:55.471210] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.925 [2024-07-15 18:12:55.471217] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.925 [2024-07-15 18:12:55.471223] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.925 [2024-07-15 18:12:55.471240] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.925 qpair failed and we were unable to recover it. 00:27:01.925 [2024-07-15 18:12:55.481192] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.925 [2024-07-15 18:12:55.481257] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.925 [2024-07-15 18:12:55.481273] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.925 [2024-07-15 18:12:55.481280] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.925 [2024-07-15 18:12:55.481286] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.925 [2024-07-15 18:12:55.481300] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.925 qpair failed and we were unable to recover it. 00:27:01.925 [2024-07-15 18:12:55.491147] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.925 [2024-07-15 18:12:55.491209] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.925 [2024-07-15 18:12:55.491227] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.925 [2024-07-15 18:12:55.491234] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.925 [2024-07-15 18:12:55.491240] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.925 [2024-07-15 18:12:55.491254] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.925 qpair failed and we were unable to recover it. 00:27:01.925 [2024-07-15 18:12:55.501210] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.925 [2024-07-15 18:12:55.501272] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.925 [2024-07-15 18:12:55.501288] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.925 [2024-07-15 18:12:55.501295] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.925 [2024-07-15 18:12:55.501300] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.925 [2024-07-15 18:12:55.501314] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.925 qpair failed and we were unable to recover it. 00:27:01.925 [2024-07-15 18:12:55.511238] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.925 [2024-07-15 18:12:55.511309] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.925 [2024-07-15 18:12:55.511324] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.925 [2024-07-15 18:12:55.511331] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.925 [2024-07-15 18:12:55.511337] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.926 [2024-07-15 18:12:55.511350] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.926 qpair failed and we were unable to recover it. 00:27:01.926 [2024-07-15 18:12:55.521270] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.926 [2024-07-15 18:12:55.521334] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.926 [2024-07-15 18:12:55.521350] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.926 [2024-07-15 18:12:55.521357] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.926 [2024-07-15 18:12:55.521362] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.926 [2024-07-15 18:12:55.521377] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.926 qpair failed and we were unable to recover it. 00:27:01.926 [2024-07-15 18:12:55.531343] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.926 [2024-07-15 18:12:55.531409] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.926 [2024-07-15 18:12:55.531423] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.926 [2024-07-15 18:12:55.531430] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.926 [2024-07-15 18:12:55.531436] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.926 [2024-07-15 18:12:55.531450] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.926 qpair failed and we were unable to recover it. 00:27:01.926 [2024-07-15 18:12:55.541253] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.926 [2024-07-15 18:12:55.541314] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.926 [2024-07-15 18:12:55.541329] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.926 [2024-07-15 18:12:55.541336] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.926 [2024-07-15 18:12:55.541342] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.926 [2024-07-15 18:12:55.541356] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.926 qpair failed and we were unable to recover it. 00:27:01.926 [2024-07-15 18:12:55.551362] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.926 [2024-07-15 18:12:55.551430] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.926 [2024-07-15 18:12:55.551446] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.926 [2024-07-15 18:12:55.551456] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.926 [2024-07-15 18:12:55.551462] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x1931ed0 00:27:01.926 [2024-07-15 18:12:55.551476] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 3 00:27:01.926 qpair failed and we were unable to recover it. 00:27:01.926 [2024-07-15 18:12:55.561392] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.926 [2024-07-15 18:12:55.561472] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.926 [2024-07-15 18:12:55.561498] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.926 [2024-07-15 18:12:55.561509] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.926 [2024-07-15 18:12:55.561518] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c0000b90 00:27:01.926 [2024-07-15 18:12:55.561541] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 4 00:27:01.926 qpair failed and we were unable to recover it. 00:27:01.926 [2024-07-15 18:12:55.571434] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.926 [2024-07-15 18:12:55.571495] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.926 [2024-07-15 18:12:55.571511] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.926 [2024-07-15 18:12:55.571518] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.926 [2024-07-15 18:12:55.571524] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c0000b90 00:27:01.926 [2024-07-15 18:12:55.571539] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 4 00:27:01.926 qpair failed and we were unable to recover it. 00:27:01.926 [2024-07-15 18:12:55.581451] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.926 [2024-07-15 18:12:55.581517] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.926 [2024-07-15 18:12:55.581538] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.926 [2024-07-15 18:12:55.581546] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.926 [2024-07-15 18:12:55.581553] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:27:01.926 [2024-07-15 18:12:55.581571] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:27:01.926 qpair failed and we were unable to recover it. 00:27:01.926 [2024-07-15 18:12:55.591459] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.926 [2024-07-15 18:12:55.591524] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.926 [2024-07-15 18:12:55.591540] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.926 [2024-07-15 18:12:55.591546] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.926 [2024-07-15 18:12:55.591552] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8c8000b90 00:27:01.926 [2024-07-15 18:12:55.591568] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 2 00:27:01.926 qpair failed and we were unable to recover it. 00:27:01.926 [2024-07-15 18:12:55.591718] nvme_ctrlr.c:4476:nvme_ctrlr_keep_alive: *ERROR*: [nqn.2016-06.io.spdk:cnode1] Submitting Keep Alive failed 00:27:01.926 A controller has encountered a failure and is being reset. 00:27:01.926 [2024-07-15 18:12:55.601507] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.926 [2024-07-15 18:12:55.601583] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.926 [2024-07-15 18:12:55.601610] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.926 [2024-07-15 18:12:55.601620] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.926 [2024-07-15 18:12:55.601629] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8d0000b90 00:27:01.926 [2024-07-15 18:12:55.601652] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:27:01.926 qpair failed and we were unable to recover it. 00:27:01.926 [2024-07-15 18:12:55.611484] ctrlr.c: 761:_nvmf_ctrlr_add_io_qpair: *ERROR*: Unknown controller ID 0x1 00:27:01.926 [2024-07-15 18:12:55.611550] nvme_fabric.c: 600:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command failed, rc -5, trtype:TCP adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1 00:27:01.926 [2024-07-15 18:12:55.611567] nvme_fabric.c: 611:_nvme_fabric_qpair_connect_poll: *ERROR*: Connect command completed with error: sct 1, sc 130 00:27:01.926 [2024-07-15 18:12:55.611574] nvme_tcp.c:2435:nvme_tcp_ctrlr_connect_qpair_poll: *ERROR*: Failed to poll NVMe-oF Fabric CONNECT command 00:27:01.926 [2024-07-15 18:12:55.611579] nvme_tcp.c:2225:nvme_tcp_qpair_process_completions: *ERROR*: Failed to connect tqpair=0x7fc8d0000b90 00:27:01.926 [2024-07-15 18:12:55.611595] nvme_qpair.c: 804:spdk_nvme_qpair_process_completions: *ERROR*: CQ transport error -6 (No such device or address) on qpair id 1 00:27:01.926 qpair failed and we were unable to recover it. 00:27:01.926 [2024-07-15 18:12:55.611710] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1940000 (9): Bad file descriptor 00:27:02.186 Controller properly reset. 00:27:02.186 Initializing NVMe Controllers 00:27:02.186 Attaching to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:27:02.186 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:cnode1 00:27:02.186 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 0 00:27:02.186 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 1 00:27:02.186 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 2 00:27:02.186 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:cnode1) with lcore 3 00:27:02.186 Initialization complete. Launching workers. 00:27:02.186 Starting thread on core 1 00:27:02.186 Starting thread on core 2 00:27:02.186 Starting thread on core 3 00:27:02.186 Starting thread on core 0 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- host/target_disconnect.sh@51 -- # sync 00:27:02.186 00:27:02.186 real 0m11.319s 00:27:02.186 user 0m21.465s 00:27:02.186 sys 0m4.266s 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@1124 -- # xtrace_disable 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect.nvmf_target_disconnect_tc2 -- common/autotest_common.sh@10 -- # set +x 00:27:02.186 ************************************ 00:27:02.186 END TEST nvmf_target_disconnect_tc2 00:27:02.186 ************************************ 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@1142 -- # return 0 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- host/target_disconnect.sh@72 -- # '[' -n '' ']' 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- host/target_disconnect.sh@76 -- # trap - SIGINT SIGTERM EXIT 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- host/target_disconnect.sh@77 -- # nvmftestfini 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@488 -- # nvmfcleanup 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@117 -- # sync 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@120 -- # set +e 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@121 -- # for i in {1..20} 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:27:02.186 rmmod nvme_tcp 00:27:02.186 rmmod nvme_fabrics 00:27:02.186 rmmod nvme_keyring 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@124 -- # set -e 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@125 -- # return 0 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@489 -- # '[' -n 746783 ']' 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@490 -- # killprocess 746783 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@948 -- # '[' -z 746783 ']' 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@952 -- # kill -0 746783 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@953 -- # uname 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 746783 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@954 -- # process_name=reactor_4 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@958 -- # '[' reactor_4 = sudo ']' 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@966 -- # echo 'killing process with pid 746783' 00:27:02.186 killing process with pid 746783 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@967 -- # kill 746783 00:27:02.186 18:12:55 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@972 -- # wait 746783 00:27:02.445 18:12:56 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:27:02.445 18:12:56 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:27:02.445 18:12:56 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:27:02.445 18:12:56 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:27:02.445 18:12:56 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@278 -- # remove_spdk_ns 00:27:02.445 18:12:56 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:02.445 18:12:56 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 14> /dev/null' 00:27:02.445 18:12:56 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:04.980 18:12:58 nvmf_tcp.nvmf_target_disconnect -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:27:04.980 00:27:04.980 real 0m19.222s 00:27:04.980 user 0m48.654s 00:27:04.981 sys 0m8.548s 00:27:04.981 18:12:58 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@1124 -- # xtrace_disable 00:27:04.981 18:12:58 nvmf_tcp.nvmf_target_disconnect -- common/autotest_common.sh@10 -- # set +x 00:27:04.981 ************************************ 00:27:04.981 END TEST nvmf_target_disconnect 00:27:04.981 ************************************ 00:27:04.981 18:12:58 nvmf_tcp -- common/autotest_common.sh@1142 -- # return 0 00:27:04.981 18:12:58 nvmf_tcp -- nvmf/nvmf.sh@126 -- # timing_exit host 00:27:04.981 18:12:58 nvmf_tcp -- common/autotest_common.sh@728 -- # xtrace_disable 00:27:04.981 18:12:58 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:27:04.981 18:12:58 nvmf_tcp -- nvmf/nvmf.sh@128 -- # trap - SIGINT SIGTERM EXIT 00:27:04.981 00:27:04.981 real 20m52.775s 00:27:04.981 user 45m14.832s 00:27:04.981 sys 6m15.830s 00:27:04.981 18:12:58 nvmf_tcp -- common/autotest_common.sh@1124 -- # xtrace_disable 00:27:04.981 18:12:58 nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:27:04.981 ************************************ 00:27:04.981 END TEST nvmf_tcp 00:27:04.981 ************************************ 00:27:04.981 18:12:58 -- common/autotest_common.sh@1142 -- # return 0 00:27:04.981 18:12:58 -- spdk/autotest.sh@288 -- # [[ 0 -eq 0 ]] 00:27:04.981 18:12:58 -- spdk/autotest.sh@289 -- # run_test spdkcli_nvmf_tcp /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/nvmf.sh --transport=tcp 00:27:04.981 18:12:58 -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:27:04.981 18:12:58 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:27:04.981 18:12:58 -- common/autotest_common.sh@10 -- # set +x 00:27:04.981 ************************************ 00:27:04.981 START TEST spdkcli_nvmf_tcp 00:27:04.981 ************************************ 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/nvmf.sh --transport=tcp 00:27:04.981 * Looking for test storage... 00:27:04.981 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/common.sh 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- spdkcli/common.sh@6 -- # spdkcli_job=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- spdkcli/common.sh@7 -- # spdk_clear_config_py=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/json_config/clear_config.py 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@7 -- # uname -s 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- paths/export.sh@5 -- # export PATH 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@47 -- # : 0 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- nvmf/common.sh@51 -- # have_pci_nics=0 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@12 -- # MATCH_FILE=spdkcli_nvmf.test 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@13 -- # SPDKCLI_BRANCH=/nvmf 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@15 -- # trap cleanup EXIT 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@17 -- # timing_enter run_nvmf_tgt 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- common/autotest_common.sh@722 -- # xtrace_disable 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@18 -- # run_nvmf_tgt 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- spdkcli/common.sh@33 -- # nvmf_tgt_pid=748315 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- spdkcli/common.sh@34 -- # waitforlisten 748315 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- common/autotest_common.sh@829 -- # '[' -z 748315 ']' 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- common/autotest_common.sh@834 -- # local max_retries=100 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:04.981 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- common/autotest_common.sh@838 -- # xtrace_disable 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:27:04.981 18:12:58 spdkcli_nvmf_tcp -- spdkcli/common.sh@32 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -m 0x3 -p 0 00:27:04.981 [2024-07-15 18:12:58.446165] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:27:04.981 [2024-07-15 18:12:58.446213] [ DPDK EAL parameters: nvmf --no-shconf -c 0x3 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid748315 ] 00:27:04.981 EAL: No free 2048 kB hugepages reported on node 1 00:27:04.981 [2024-07-15 18:12:58.501596] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 2 00:27:04.981 [2024-07-15 18:12:58.582595] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:27:04.981 [2024-07-15 18:12:58.582598] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:27:05.549 18:12:59 spdkcli_nvmf_tcp -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:27:05.549 18:12:59 spdkcli_nvmf_tcp -- common/autotest_common.sh@862 -- # return 0 00:27:05.549 18:12:59 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@19 -- # timing_exit run_nvmf_tgt 00:27:05.549 18:12:59 spdkcli_nvmf_tcp -- common/autotest_common.sh@728 -- # xtrace_disable 00:27:05.549 18:12:59 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:27:05.808 18:12:59 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@21 -- # NVMF_TARGET_IP=127.0.0.1 00:27:05.808 18:12:59 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@22 -- # [[ tcp == \r\d\m\a ]] 00:27:05.808 18:12:59 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@27 -- # timing_enter spdkcli_create_nvmf_config 00:27:05.808 18:12:59 spdkcli_nvmf_tcp -- common/autotest_common.sh@722 -- # xtrace_disable 00:27:05.808 18:12:59 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:27:05.808 18:12:59 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@65 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py ''\''/bdevs/malloc create 32 512 Malloc1'\'' '\''Malloc1'\'' True 00:27:05.808 '\''/bdevs/malloc create 32 512 Malloc2'\'' '\''Malloc2'\'' True 00:27:05.808 '\''/bdevs/malloc create 32 512 Malloc3'\'' '\''Malloc3'\'' True 00:27:05.808 '\''/bdevs/malloc create 32 512 Malloc4'\'' '\''Malloc4'\'' True 00:27:05.808 '\''/bdevs/malloc create 32 512 Malloc5'\'' '\''Malloc5'\'' True 00:27:05.808 '\''/bdevs/malloc create 32 512 Malloc6'\'' '\''Malloc6'\'' True 00:27:05.808 '\''nvmf/transport create tcp max_io_qpairs_per_ctrlr=4 io_unit_size=8192'\'' '\'''\'' True 00:27:05.808 '\''/nvmf/subsystem create nqn.2014-08.org.spdk:cnode1 N37SXV509SRW max_namespaces=4 allow_any_host=True'\'' '\''nqn.2014-08.org.spdk:cnode1'\'' True 00:27:05.808 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc3 1'\'' '\''Malloc3'\'' True 00:27:05.808 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc4 2'\'' '\''Malloc4'\'' True 00:27:05.808 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4260 IPv4'\'' '\''127.0.0.1:4260'\'' True 00:27:05.808 '\''/nvmf/subsystem create nqn.2014-08.org.spdk:cnode2 N37SXV509SRD max_namespaces=2 allow_any_host=True'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:27:05.808 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/namespaces create Malloc2'\'' '\''Malloc2'\'' True 00:27:05.808 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/listen_addresses create tcp 127.0.0.1 4260 IPv4'\'' '\''127.0.0.1:4260'\'' True 00:27:05.808 '\''/nvmf/subsystem create nqn.2014-08.org.spdk:cnode3 N37SXV509SRR max_namespaces=2 allow_any_host=True'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:27:05.808 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/namespaces create Malloc1'\'' '\''Malloc1'\'' True 00:27:05.808 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4260 IPv4'\'' '\''127.0.0.1:4260'\'' True 00:27:05.808 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4261 IPv4'\'' '\''127.0.0.1:4261'\'' True 00:27:05.808 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode1'\'' '\''nqn.2014-08.org.spdk:cnode1'\'' True 00:27:05.808 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode2'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:27:05.808 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host True'\'' '\''Allow any host'\'' 00:27:05.808 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host False'\'' '\''Allow any host'\'' True 00:27:05.808 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4261 IPv4'\'' '\''127.0.0.1:4261'\'' True 00:27:05.808 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4262 IPv4'\'' '\''127.0.0.1:4262'\'' True 00:27:05.808 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts create nqn.2014-08.org.spdk:cnode2'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' True 00:27:05.808 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc5'\'' '\''Malloc5'\'' True 00:27:05.808 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc6'\'' '\''Malloc6'\'' True 00:27:05.808 '\''/nvmf/referral create tcp 127.0.0.2 4030 IPv4'\'' 00:27:05.808 ' 00:27:08.341 [2024-07-15 18:13:01.663297] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:27:09.278 [2024-07-15 18:13:02.839167] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4260 *** 00:27:11.812 [2024-07-15 18:13:05.001879] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4261 *** 00:27:13.189 [2024-07-15 18:13:06.863677] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4262 *** 00:27:14.565 Executing command: ['/bdevs/malloc create 32 512 Malloc1', 'Malloc1', True] 00:27:14.565 Executing command: ['/bdevs/malloc create 32 512 Malloc2', 'Malloc2', True] 00:27:14.565 Executing command: ['/bdevs/malloc create 32 512 Malloc3', 'Malloc3', True] 00:27:14.565 Executing command: ['/bdevs/malloc create 32 512 Malloc4', 'Malloc4', True] 00:27:14.565 Executing command: ['/bdevs/malloc create 32 512 Malloc5', 'Malloc5', True] 00:27:14.565 Executing command: ['/bdevs/malloc create 32 512 Malloc6', 'Malloc6', True] 00:27:14.565 Executing command: ['nvmf/transport create tcp max_io_qpairs_per_ctrlr=4 io_unit_size=8192', '', True] 00:27:14.565 Executing command: ['/nvmf/subsystem create nqn.2014-08.org.spdk:cnode1 N37SXV509SRW max_namespaces=4 allow_any_host=True', 'nqn.2014-08.org.spdk:cnode1', True] 00:27:14.565 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc3 1', 'Malloc3', True] 00:27:14.565 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc4 2', 'Malloc4', True] 00:27:14.565 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4260 IPv4', '127.0.0.1:4260', True] 00:27:14.565 Executing command: ['/nvmf/subsystem create nqn.2014-08.org.spdk:cnode2 N37SXV509SRD max_namespaces=2 allow_any_host=True', 'nqn.2014-08.org.spdk:cnode2', True] 00:27:14.565 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/namespaces create Malloc2', 'Malloc2', True] 00:27:14.565 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode2/listen_addresses create tcp 127.0.0.1 4260 IPv4', '127.0.0.1:4260', True] 00:27:14.565 Executing command: ['/nvmf/subsystem create nqn.2014-08.org.spdk:cnode3 N37SXV509SRR max_namespaces=2 allow_any_host=True', 'nqn.2014-08.org.spdk:cnode2', True] 00:27:14.565 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/namespaces create Malloc1', 'Malloc1', True] 00:27:14.565 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4260 IPv4', '127.0.0.1:4260', True] 00:27:14.565 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/listen_addresses create tcp 127.0.0.1 4261 IPv4', '127.0.0.1:4261', True] 00:27:14.565 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode1', 'nqn.2014-08.org.spdk:cnode1', True] 00:27:14.565 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts create nqn.2014-08.org.spdk:cnode2', 'nqn.2014-08.org.spdk:cnode2', True] 00:27:14.565 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host True', 'Allow any host', False] 00:27:14.565 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1 allow_any_host False', 'Allow any host', True] 00:27:14.565 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4261 IPv4', '127.0.0.1:4261', True] 00:27:14.565 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses create tcp 127.0.0.1 4262 IPv4', '127.0.0.1:4262', True] 00:27:14.565 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts create nqn.2014-08.org.spdk:cnode2', 'nqn.2014-08.org.spdk:cnode2', True] 00:27:14.565 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc5', 'Malloc5', True] 00:27:14.565 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces create Malloc6', 'Malloc6', True] 00:27:14.565 Executing command: ['/nvmf/referral create tcp 127.0.0.2 4030 IPv4', False] 00:27:14.822 18:13:08 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@66 -- # timing_exit spdkcli_create_nvmf_config 00:27:14.822 18:13:08 spdkcli_nvmf_tcp -- common/autotest_common.sh@728 -- # xtrace_disable 00:27:14.822 18:13:08 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:27:14.822 18:13:08 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@68 -- # timing_enter spdkcli_check_match 00:27:14.822 18:13:08 spdkcli_nvmf_tcp -- common/autotest_common.sh@722 -- # xtrace_disable 00:27:14.822 18:13:08 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:27:14.822 18:13:08 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@69 -- # check_match 00:27:14.822 18:13:08 spdkcli_nvmf_tcp -- spdkcli/common.sh@44 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdkcli.py ll /nvmf 00:27:15.080 18:13:08 spdkcli_nvmf_tcp -- spdkcli/common.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/app/match/match /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/match_files/spdkcli_nvmf.test.match 00:27:15.338 18:13:08 spdkcli_nvmf_tcp -- spdkcli/common.sh@46 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/match_files/spdkcli_nvmf.test 00:27:15.338 18:13:08 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@70 -- # timing_exit spdkcli_check_match 00:27:15.338 18:13:08 spdkcli_nvmf_tcp -- common/autotest_common.sh@728 -- # xtrace_disable 00:27:15.338 18:13:08 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:27:15.338 18:13:08 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@72 -- # timing_enter spdkcli_clear_nvmf_config 00:27:15.338 18:13:08 spdkcli_nvmf_tcp -- common/autotest_common.sh@722 -- # xtrace_disable 00:27:15.338 18:13:08 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:27:15.338 18:13:08 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@87 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_job.py ''\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete nsid=1'\'' '\''Malloc3'\'' 00:27:15.338 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete_all'\'' '\''Malloc4'\'' 00:27:15.338 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts delete nqn.2014-08.org.spdk:cnode2'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' 00:27:15.338 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts delete_all'\'' '\''nqn.2014-08.org.spdk:cnode1'\'' 00:27:15.338 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete tcp 127.0.0.1 4262'\'' '\''127.0.0.1:4262'\'' 00:27:15.338 '\''/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete_all'\'' '\''127.0.0.1:4261'\'' 00:27:15.338 '\''/nvmf/subsystem delete nqn.2014-08.org.spdk:cnode3'\'' '\''nqn.2014-08.org.spdk:cnode3'\'' 00:27:15.338 '\''/nvmf/subsystem delete_all'\'' '\''nqn.2014-08.org.spdk:cnode2'\'' 00:27:15.338 '\''/bdevs/malloc delete Malloc6'\'' '\''Malloc6'\'' 00:27:15.338 '\''/bdevs/malloc delete Malloc5'\'' '\''Malloc5'\'' 00:27:15.338 '\''/bdevs/malloc delete Malloc4'\'' '\''Malloc4'\'' 00:27:15.338 '\''/bdevs/malloc delete Malloc3'\'' '\''Malloc3'\'' 00:27:15.338 '\''/bdevs/malloc delete Malloc2'\'' '\''Malloc2'\'' 00:27:15.338 '\''/bdevs/malloc delete Malloc1'\'' '\''Malloc1'\'' 00:27:15.338 ' 00:27:20.728 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete nsid=1', 'Malloc3', False] 00:27:20.728 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/namespaces delete_all', 'Malloc4', False] 00:27:20.728 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/hosts delete nqn.2014-08.org.spdk:cnode2', 'nqn.2014-08.org.spdk:cnode2', False] 00:27:20.728 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode3/hosts delete_all', 'nqn.2014-08.org.spdk:cnode1', False] 00:27:20.728 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete tcp 127.0.0.1 4262', '127.0.0.1:4262', False] 00:27:20.728 Executing command: ['/nvmf/subsystem/nqn.2014-08.org.spdk:cnode1/listen_addresses delete_all', '127.0.0.1:4261', False] 00:27:20.728 Executing command: ['/nvmf/subsystem delete nqn.2014-08.org.spdk:cnode3', 'nqn.2014-08.org.spdk:cnode3', False] 00:27:20.728 Executing command: ['/nvmf/subsystem delete_all', 'nqn.2014-08.org.spdk:cnode2', False] 00:27:20.728 Executing command: ['/bdevs/malloc delete Malloc6', 'Malloc6', False] 00:27:20.728 Executing command: ['/bdevs/malloc delete Malloc5', 'Malloc5', False] 00:27:20.728 Executing command: ['/bdevs/malloc delete Malloc4', 'Malloc4', False] 00:27:20.728 Executing command: ['/bdevs/malloc delete Malloc3', 'Malloc3', False] 00:27:20.728 Executing command: ['/bdevs/malloc delete Malloc2', 'Malloc2', False] 00:27:20.728 Executing command: ['/bdevs/malloc delete Malloc1', 'Malloc1', False] 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@88 -- # timing_exit spdkcli_clear_nvmf_config 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- common/autotest_common.sh@728 -- # xtrace_disable 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@90 -- # killprocess 748315 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- common/autotest_common.sh@948 -- # '[' -z 748315 ']' 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- common/autotest_common.sh@952 -- # kill -0 748315 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- common/autotest_common.sh@953 -- # uname 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 748315 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- common/autotest_common.sh@966 -- # echo 'killing process with pid 748315' 00:27:20.728 killing process with pid 748315 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- common/autotest_common.sh@967 -- # kill 748315 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- common/autotest_common.sh@972 -- # wait 748315 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- spdkcli/nvmf.sh@1 -- # cleanup 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- spdkcli/common.sh@10 -- # '[' -n '' ']' 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- spdkcli/common.sh@13 -- # '[' -n 748315 ']' 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- spdkcli/common.sh@14 -- # killprocess 748315 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- common/autotest_common.sh@948 -- # '[' -z 748315 ']' 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- common/autotest_common.sh@952 -- # kill -0 748315 00:27:20.728 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 952: kill: (748315) - No such process 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- common/autotest_common.sh@975 -- # echo 'Process with pid 748315 is not found' 00:27:20.728 Process with pid 748315 is not found 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- spdkcli/common.sh@16 -- # '[' -n '' ']' 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- spdkcli/common.sh@19 -- # '[' -n '' ']' 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- spdkcli/common.sh@22 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/spdkcli_nvmf.test /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/spdkcli/match_files/spdkcli_details_vhost.test /tmp/sample_aio 00:27:20.728 00:27:20.728 real 0m16.110s 00:27:20.728 user 0m33.885s 00:27:20.728 sys 0m0.691s 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- common/autotest_common.sh@1124 -- # xtrace_disable 00:27:20.728 18:13:14 spdkcli_nvmf_tcp -- common/autotest_common.sh@10 -- # set +x 00:27:20.728 ************************************ 00:27:20.728 END TEST spdkcli_nvmf_tcp 00:27:20.728 ************************************ 00:27:20.728 18:13:14 -- common/autotest_common.sh@1142 -- # return 0 00:27:20.728 18:13:14 -- spdk/autotest.sh@290 -- # run_test nvmf_identify_passthru /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/identify_passthru.sh --transport=tcp 00:27:20.728 18:13:14 -- common/autotest_common.sh@1099 -- # '[' 3 -le 1 ']' 00:27:20.728 18:13:14 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:27:20.728 18:13:14 -- common/autotest_common.sh@10 -- # set +x 00:27:20.988 ************************************ 00:27:20.988 START TEST nvmf_identify_passthru 00:27:20.988 ************************************ 00:27:20.988 18:13:14 nvmf_identify_passthru -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/identify_passthru.sh --transport=tcp 00:27:20.988 * Looking for test storage... 00:27:20.988 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:27:20.988 18:13:14 nvmf_identify_passthru -- target/identify_passthru.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@7 -- # uname -s 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:27:20.988 18:13:14 nvmf_identify_passthru -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:27:20.988 18:13:14 nvmf_identify_passthru -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:27:20.988 18:13:14 nvmf_identify_passthru -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:27:20.988 18:13:14 nvmf_identify_passthru -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:20.988 18:13:14 nvmf_identify_passthru -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:20.988 18:13:14 nvmf_identify_passthru -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:20.988 18:13:14 nvmf_identify_passthru -- paths/export.sh@5 -- # export PATH 00:27:20.988 18:13:14 nvmf_identify_passthru -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@47 -- # : 0 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@51 -- # have_pci_nics=0 00:27:20.988 18:13:14 nvmf_identify_passthru -- target/identify_passthru.sh@10 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:27:20.988 18:13:14 nvmf_identify_passthru -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:27:20.988 18:13:14 nvmf_identify_passthru -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:27:20.988 18:13:14 nvmf_identify_passthru -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:27:20.988 18:13:14 nvmf_identify_passthru -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:20.988 18:13:14 nvmf_identify_passthru -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:20.988 18:13:14 nvmf_identify_passthru -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:20.988 18:13:14 nvmf_identify_passthru -- paths/export.sh@5 -- # export PATH 00:27:20.988 18:13:14 nvmf_identify_passthru -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:20.988 18:13:14 nvmf_identify_passthru -- target/identify_passthru.sh@12 -- # nvmftestinit 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@448 -- # prepare_net_devs 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@410 -- # local -g is_hw=no 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@412 -- # remove_spdk_ns 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:20.988 18:13:14 nvmf_identify_passthru -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:27:20.988 18:13:14 nvmf_identify_passthru -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:27:20.988 18:13:14 nvmf_identify_passthru -- nvmf/common.sh@285 -- # xtrace_disable 00:27:20.988 18:13:14 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@291 -- # pci_devs=() 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@291 -- # local -a pci_devs 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@292 -- # pci_net_devs=() 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@293 -- # pci_drivers=() 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@293 -- # local -A pci_drivers 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@295 -- # net_devs=() 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@295 -- # local -ga net_devs 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@296 -- # e810=() 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@296 -- # local -ga e810 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@297 -- # x722=() 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@297 -- # local -ga x722 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@298 -- # mlx=() 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@298 -- # local -ga mlx 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:27:26.258 Found 0000:86:00.0 (0x8086 - 0x159b) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:27:26.258 Found 0000:86:00.1 (0x8086 - 0x159b) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@390 -- # [[ up == up ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:27:26.258 Found net devices under 0000:86:00.0: cvl_0_0 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@390 -- # [[ up == up ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:27:26.258 Found net devices under 0000:86:00.1: cvl_0_1 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@414 -- # is_hw=yes 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:27:26.258 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:27:26.258 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.270 ms 00:27:26.258 00:27:26.258 --- 10.0.0.2 ping statistics --- 00:27:26.258 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:26.258 rtt min/avg/max/mdev = 0.270/0.270/0.270/0.000 ms 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:27:26.258 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:27:26.258 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.238 ms 00:27:26.258 00:27:26.258 --- 10.0.0.1 ping statistics --- 00:27:26.258 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:26.258 rtt min/avg/max/mdev = 0.238/0.238/0.238/0.000 ms 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@422 -- # return 0 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@450 -- # '[' '' == iso ']' 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:27:26.258 18:13:19 nvmf_identify_passthru -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:27:26.258 18:13:19 nvmf_identify_passthru -- target/identify_passthru.sh@14 -- # timing_enter nvme_identify 00:27:26.258 18:13:19 nvmf_identify_passthru -- common/autotest_common.sh@722 -- # xtrace_disable 00:27:26.258 18:13:19 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:27:26.258 18:13:19 nvmf_identify_passthru -- target/identify_passthru.sh@16 -- # get_first_nvme_bdf 00:27:26.258 18:13:19 nvmf_identify_passthru -- common/autotest_common.sh@1524 -- # bdfs=() 00:27:26.258 18:13:19 nvmf_identify_passthru -- common/autotest_common.sh@1524 -- # local bdfs 00:27:26.258 18:13:19 nvmf_identify_passthru -- common/autotest_common.sh@1525 -- # bdfs=($(get_nvme_bdfs)) 00:27:26.258 18:13:19 nvmf_identify_passthru -- common/autotest_common.sh@1525 -- # get_nvme_bdfs 00:27:26.258 18:13:19 nvmf_identify_passthru -- common/autotest_common.sh@1513 -- # bdfs=() 00:27:26.258 18:13:19 nvmf_identify_passthru -- common/autotest_common.sh@1513 -- # local bdfs 00:27:26.258 18:13:19 nvmf_identify_passthru -- common/autotest_common.sh@1514 -- # bdfs=($("$rootdir/scripts/gen_nvme.sh" | jq -r '.config[].params.traddr')) 00:27:26.258 18:13:19 nvmf_identify_passthru -- common/autotest_common.sh@1514 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/gen_nvme.sh 00:27:26.258 18:13:19 nvmf_identify_passthru -- common/autotest_common.sh@1514 -- # jq -r '.config[].params.traddr' 00:27:26.258 18:13:19 nvmf_identify_passthru -- common/autotest_common.sh@1515 -- # (( 1 == 0 )) 00:27:26.258 18:13:19 nvmf_identify_passthru -- common/autotest_common.sh@1519 -- # printf '%s\n' 0000:5e:00.0 00:27:26.258 18:13:19 nvmf_identify_passthru -- common/autotest_common.sh@1527 -- # echo 0000:5e:00.0 00:27:26.258 18:13:19 nvmf_identify_passthru -- target/identify_passthru.sh@16 -- # bdf=0000:5e:00.0 00:27:26.258 18:13:19 nvmf_identify_passthru -- target/identify_passthru.sh@17 -- # '[' -z 0000:5e:00.0 ']' 00:27:26.258 18:13:19 nvmf_identify_passthru -- target/identify_passthru.sh@23 -- # awk '{print $3}' 00:27:26.258 18:13:19 nvmf_identify_passthru -- target/identify_passthru.sh@23 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:PCIe traddr:0000:5e:00.0' -i 0 00:27:26.258 18:13:19 nvmf_identify_passthru -- target/identify_passthru.sh@23 -- # grep 'Serial Number:' 00:27:26.258 EAL: No free 2048 kB hugepages reported on node 1 00:27:30.446 18:13:23 nvmf_identify_passthru -- target/identify_passthru.sh@23 -- # nvme_serial_number=BTLJ72430F0E1P0FGN 00:27:30.446 18:13:23 nvmf_identify_passthru -- target/identify_passthru.sh@24 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r 'trtype:PCIe traddr:0000:5e:00.0' -i 0 00:27:30.446 18:13:23 nvmf_identify_passthru -- target/identify_passthru.sh@24 -- # grep 'Model Number:' 00:27:30.446 18:13:23 nvmf_identify_passthru -- target/identify_passthru.sh@24 -- # awk '{print $3}' 00:27:30.446 EAL: No free 2048 kB hugepages reported on node 1 00:27:34.633 18:13:28 nvmf_identify_passthru -- target/identify_passthru.sh@24 -- # nvme_model_number=INTEL 00:27:34.633 18:13:28 nvmf_identify_passthru -- target/identify_passthru.sh@26 -- # timing_exit nvme_identify 00:27:34.633 18:13:28 nvmf_identify_passthru -- common/autotest_common.sh@728 -- # xtrace_disable 00:27:34.633 18:13:28 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:27:34.633 18:13:28 nvmf_identify_passthru -- target/identify_passthru.sh@28 -- # timing_enter start_nvmf_tgt 00:27:34.633 18:13:28 nvmf_identify_passthru -- common/autotest_common.sh@722 -- # xtrace_disable 00:27:34.633 18:13:28 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:27:34.633 18:13:28 nvmf_identify_passthru -- target/identify_passthru.sh@31 -- # nvmfpid=755333 00:27:34.633 18:13:28 nvmf_identify_passthru -- target/identify_passthru.sh@30 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xF --wait-for-rpc 00:27:34.633 18:13:28 nvmf_identify_passthru -- target/identify_passthru.sh@33 -- # trap 'process_shm --id $NVMF_APP_SHM_ID; nvmftestfini; exit 1' SIGINT SIGTERM EXIT 00:27:34.633 18:13:28 nvmf_identify_passthru -- target/identify_passthru.sh@35 -- # waitforlisten 755333 00:27:34.633 18:13:28 nvmf_identify_passthru -- common/autotest_common.sh@829 -- # '[' -z 755333 ']' 00:27:34.633 18:13:28 nvmf_identify_passthru -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:34.633 18:13:28 nvmf_identify_passthru -- common/autotest_common.sh@834 -- # local max_retries=100 00:27:34.633 18:13:28 nvmf_identify_passthru -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:34.633 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:34.633 18:13:28 nvmf_identify_passthru -- common/autotest_common.sh@838 -- # xtrace_disable 00:27:34.633 18:13:28 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:27:34.633 [2024-07-15 18:13:28.186187] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:27:34.634 [2024-07-15 18:13:28.186240] [ DPDK EAL parameters: nvmf -c 0xF --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:34.634 EAL: No free 2048 kB hugepages reported on node 1 00:27:34.634 [2024-07-15 18:13:28.243262] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:27:34.634 [2024-07-15 18:13:28.324132] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:27:34.634 [2024-07-15 18:13:28.324169] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:27:34.634 [2024-07-15 18:13:28.324176] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:27:34.634 [2024-07-15 18:13:28.324183] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:27:34.634 [2024-07-15 18:13:28.324188] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:27:34.634 [2024-07-15 18:13:28.324235] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:27:34.634 [2024-07-15 18:13:28.324299] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:27:34.634 [2024-07-15 18:13:28.324372] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:27:34.634 [2024-07-15 18:13:28.324374] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:27:35.568 18:13:28 nvmf_identify_passthru -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:27:35.568 18:13:29 nvmf_identify_passthru -- common/autotest_common.sh@862 -- # return 0 00:27:35.568 18:13:29 nvmf_identify_passthru -- target/identify_passthru.sh@36 -- # rpc_cmd -v nvmf_set_config --passthru-identify-ctrlr 00:27:35.568 18:13:29 nvmf_identify_passthru -- common/autotest_common.sh@559 -- # xtrace_disable 00:27:35.568 18:13:29 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:27:35.568 INFO: Log level set to 20 00:27:35.568 INFO: Requests: 00:27:35.568 { 00:27:35.568 "jsonrpc": "2.0", 00:27:35.568 "method": "nvmf_set_config", 00:27:35.568 "id": 1, 00:27:35.568 "params": { 00:27:35.568 "admin_cmd_passthru": { 00:27:35.568 "identify_ctrlr": true 00:27:35.568 } 00:27:35.568 } 00:27:35.568 } 00:27:35.568 00:27:35.568 INFO: response: 00:27:35.568 { 00:27:35.568 "jsonrpc": "2.0", 00:27:35.568 "id": 1, 00:27:35.568 "result": true 00:27:35.568 } 00:27:35.568 00:27:35.568 18:13:29 nvmf_identify_passthru -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:27:35.568 18:13:29 nvmf_identify_passthru -- target/identify_passthru.sh@37 -- # rpc_cmd -v framework_start_init 00:27:35.568 18:13:29 nvmf_identify_passthru -- common/autotest_common.sh@559 -- # xtrace_disable 00:27:35.568 18:13:29 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:27:35.568 INFO: Setting log level to 20 00:27:35.568 INFO: Setting log level to 20 00:27:35.568 INFO: Log level set to 20 00:27:35.568 INFO: Log level set to 20 00:27:35.568 INFO: Requests: 00:27:35.568 { 00:27:35.568 "jsonrpc": "2.0", 00:27:35.568 "method": "framework_start_init", 00:27:35.568 "id": 1 00:27:35.568 } 00:27:35.568 00:27:35.568 INFO: Requests: 00:27:35.568 { 00:27:35.568 "jsonrpc": "2.0", 00:27:35.568 "method": "framework_start_init", 00:27:35.568 "id": 1 00:27:35.568 } 00:27:35.568 00:27:35.568 [2024-07-15 18:13:29.094712] nvmf_tgt.c: 451:nvmf_tgt_advance_state: *NOTICE*: Custom identify ctrlr handler enabled 00:27:35.568 INFO: response: 00:27:35.568 { 00:27:35.568 "jsonrpc": "2.0", 00:27:35.568 "id": 1, 00:27:35.568 "result": true 00:27:35.568 } 00:27:35.568 00:27:35.568 INFO: response: 00:27:35.568 { 00:27:35.568 "jsonrpc": "2.0", 00:27:35.568 "id": 1, 00:27:35.568 "result": true 00:27:35.568 } 00:27:35.568 00:27:35.568 18:13:29 nvmf_identify_passthru -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:27:35.568 18:13:29 nvmf_identify_passthru -- target/identify_passthru.sh@38 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:27:35.568 18:13:29 nvmf_identify_passthru -- common/autotest_common.sh@559 -- # xtrace_disable 00:27:35.568 18:13:29 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:27:35.568 INFO: Setting log level to 40 00:27:35.568 INFO: Setting log level to 40 00:27:35.568 INFO: Setting log level to 40 00:27:35.568 [2024-07-15 18:13:29.108244] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:27:35.568 18:13:29 nvmf_identify_passthru -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:27:35.568 18:13:29 nvmf_identify_passthru -- target/identify_passthru.sh@39 -- # timing_exit start_nvmf_tgt 00:27:35.568 18:13:29 nvmf_identify_passthru -- common/autotest_common.sh@728 -- # xtrace_disable 00:27:35.568 18:13:29 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:27:35.568 18:13:29 nvmf_identify_passthru -- target/identify_passthru.sh@41 -- # rpc_cmd bdev_nvme_attach_controller -b Nvme0 -t PCIe -a 0000:5e:00.0 00:27:35.568 18:13:29 nvmf_identify_passthru -- common/autotest_common.sh@559 -- # xtrace_disable 00:27:35.569 18:13:29 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:27:38.856 Nvme0n1 00:27:38.856 18:13:31 nvmf_identify_passthru -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:27:38.856 18:13:31 nvmf_identify_passthru -- target/identify_passthru.sh@42 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 -a -s SPDK00000000000001 -m 1 00:27:38.856 18:13:31 nvmf_identify_passthru -- common/autotest_common.sh@559 -- # xtrace_disable 00:27:38.856 18:13:31 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:27:38.856 18:13:31 nvmf_identify_passthru -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:27:38.856 18:13:31 nvmf_identify_passthru -- target/identify_passthru.sh@43 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 Nvme0n1 00:27:38.856 18:13:31 nvmf_identify_passthru -- common/autotest_common.sh@559 -- # xtrace_disable 00:27:38.856 18:13:31 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:27:38.856 18:13:31 nvmf_identify_passthru -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:27:38.856 18:13:31 nvmf_identify_passthru -- target/identify_passthru.sh@44 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:27:38.856 18:13:31 nvmf_identify_passthru -- common/autotest_common.sh@559 -- # xtrace_disable 00:27:38.856 18:13:31 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:27:38.856 [2024-07-15 18:13:32.000503] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:27:38.857 18:13:32 nvmf_identify_passthru -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:27:38.857 18:13:32 nvmf_identify_passthru -- target/identify_passthru.sh@46 -- # rpc_cmd nvmf_get_subsystems 00:27:38.857 18:13:32 nvmf_identify_passthru -- common/autotest_common.sh@559 -- # xtrace_disable 00:27:38.857 18:13:32 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:27:38.857 [ 00:27:38.857 { 00:27:38.857 "nqn": "nqn.2014-08.org.nvmexpress.discovery", 00:27:38.857 "subtype": "Discovery", 00:27:38.857 "listen_addresses": [], 00:27:38.857 "allow_any_host": true, 00:27:38.857 "hosts": [] 00:27:38.857 }, 00:27:38.857 { 00:27:38.857 "nqn": "nqn.2016-06.io.spdk:cnode1", 00:27:38.857 "subtype": "NVMe", 00:27:38.857 "listen_addresses": [ 00:27:38.857 { 00:27:38.857 "trtype": "TCP", 00:27:38.857 "adrfam": "IPv4", 00:27:38.857 "traddr": "10.0.0.2", 00:27:38.857 "trsvcid": "4420" 00:27:38.857 } 00:27:38.857 ], 00:27:38.857 "allow_any_host": true, 00:27:38.857 "hosts": [], 00:27:38.857 "serial_number": "SPDK00000000000001", 00:27:38.857 "model_number": "SPDK bdev Controller", 00:27:38.857 "max_namespaces": 1, 00:27:38.857 "min_cntlid": 1, 00:27:38.857 "max_cntlid": 65519, 00:27:38.857 "namespaces": [ 00:27:38.857 { 00:27:38.857 "nsid": 1, 00:27:38.857 "bdev_name": "Nvme0n1", 00:27:38.857 "name": "Nvme0n1", 00:27:38.857 "nguid": "061528B45FCB474B8D3D94105E0AF548", 00:27:38.857 "uuid": "061528b4-5fcb-474b-8d3d-94105e0af548" 00:27:38.857 } 00:27:38.857 ] 00:27:38.857 } 00:27:38.857 ] 00:27:38.857 18:13:32 nvmf_identify_passthru -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:27:38.857 18:13:32 nvmf_identify_passthru -- target/identify_passthru.sh@54 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:27:38.857 18:13:32 nvmf_identify_passthru -- target/identify_passthru.sh@54 -- # grep 'Serial Number:' 00:27:38.857 18:13:32 nvmf_identify_passthru -- target/identify_passthru.sh@54 -- # awk '{print $3}' 00:27:38.857 EAL: No free 2048 kB hugepages reported on node 1 00:27:38.857 18:13:32 nvmf_identify_passthru -- target/identify_passthru.sh@54 -- # nvmf_serial_number=BTLJ72430F0E1P0FGN 00:27:38.857 18:13:32 nvmf_identify_passthru -- target/identify_passthru.sh@61 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_nvme_identify -r ' trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:cnode1' 00:27:38.857 18:13:32 nvmf_identify_passthru -- target/identify_passthru.sh@61 -- # grep 'Model Number:' 00:27:38.857 18:13:32 nvmf_identify_passthru -- target/identify_passthru.sh@61 -- # awk '{print $3}' 00:27:38.857 EAL: No free 2048 kB hugepages reported on node 1 00:27:38.857 18:13:32 nvmf_identify_passthru -- target/identify_passthru.sh@61 -- # nvmf_model_number=INTEL 00:27:38.857 18:13:32 nvmf_identify_passthru -- target/identify_passthru.sh@63 -- # '[' BTLJ72430F0E1P0FGN '!=' BTLJ72430F0E1P0FGN ']' 00:27:38.857 18:13:32 nvmf_identify_passthru -- target/identify_passthru.sh@68 -- # '[' INTEL '!=' INTEL ']' 00:27:38.857 18:13:32 nvmf_identify_passthru -- target/identify_passthru.sh@73 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:27:38.857 18:13:32 nvmf_identify_passthru -- common/autotest_common.sh@559 -- # xtrace_disable 00:27:38.857 18:13:32 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:27:38.857 18:13:32 nvmf_identify_passthru -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:27:38.857 18:13:32 nvmf_identify_passthru -- target/identify_passthru.sh@75 -- # trap - SIGINT SIGTERM EXIT 00:27:38.857 18:13:32 nvmf_identify_passthru -- target/identify_passthru.sh@77 -- # nvmftestfini 00:27:38.857 18:13:32 nvmf_identify_passthru -- nvmf/common.sh@488 -- # nvmfcleanup 00:27:38.857 18:13:32 nvmf_identify_passthru -- nvmf/common.sh@117 -- # sync 00:27:38.857 18:13:32 nvmf_identify_passthru -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:27:38.857 18:13:32 nvmf_identify_passthru -- nvmf/common.sh@120 -- # set +e 00:27:38.857 18:13:32 nvmf_identify_passthru -- nvmf/common.sh@121 -- # for i in {1..20} 00:27:38.857 18:13:32 nvmf_identify_passthru -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:27:38.857 rmmod nvme_tcp 00:27:38.857 rmmod nvme_fabrics 00:27:38.857 rmmod nvme_keyring 00:27:38.857 18:13:32 nvmf_identify_passthru -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:27:38.857 18:13:32 nvmf_identify_passthru -- nvmf/common.sh@124 -- # set -e 00:27:38.857 18:13:32 nvmf_identify_passthru -- nvmf/common.sh@125 -- # return 0 00:27:38.857 18:13:32 nvmf_identify_passthru -- nvmf/common.sh@489 -- # '[' -n 755333 ']' 00:27:38.857 18:13:32 nvmf_identify_passthru -- nvmf/common.sh@490 -- # killprocess 755333 00:27:38.857 18:13:32 nvmf_identify_passthru -- common/autotest_common.sh@948 -- # '[' -z 755333 ']' 00:27:38.857 18:13:32 nvmf_identify_passthru -- common/autotest_common.sh@952 -- # kill -0 755333 00:27:38.857 18:13:32 nvmf_identify_passthru -- common/autotest_common.sh@953 -- # uname 00:27:38.857 18:13:32 nvmf_identify_passthru -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:27:38.857 18:13:32 nvmf_identify_passthru -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 755333 00:27:38.857 18:13:32 nvmf_identify_passthru -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:27:38.857 18:13:32 nvmf_identify_passthru -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:27:38.857 18:13:32 nvmf_identify_passthru -- common/autotest_common.sh@966 -- # echo 'killing process with pid 755333' 00:27:38.857 killing process with pid 755333 00:27:38.857 18:13:32 nvmf_identify_passthru -- common/autotest_common.sh@967 -- # kill 755333 00:27:38.857 18:13:32 nvmf_identify_passthru -- common/autotest_common.sh@972 -- # wait 755333 00:27:40.249 18:13:33 nvmf_identify_passthru -- nvmf/common.sh@492 -- # '[' '' == iso ']' 00:27:40.249 18:13:33 nvmf_identify_passthru -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:27:40.249 18:13:33 nvmf_identify_passthru -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:27:40.249 18:13:33 nvmf_identify_passthru -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:27:40.249 18:13:33 nvmf_identify_passthru -- nvmf/common.sh@278 -- # remove_spdk_ns 00:27:40.249 18:13:33 nvmf_identify_passthru -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:40.249 18:13:33 nvmf_identify_passthru -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:27:40.249 18:13:33 nvmf_identify_passthru -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:42.789 18:13:35 nvmf_identify_passthru -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:27:42.789 00:27:42.789 real 0m21.481s 00:27:42.789 user 0m29.568s 00:27:42.789 sys 0m4.693s 00:27:42.789 18:13:35 nvmf_identify_passthru -- common/autotest_common.sh@1124 -- # xtrace_disable 00:27:42.789 18:13:35 nvmf_identify_passthru -- common/autotest_common.sh@10 -- # set +x 00:27:42.789 ************************************ 00:27:42.789 END TEST nvmf_identify_passthru 00:27:42.789 ************************************ 00:27:42.789 18:13:35 -- common/autotest_common.sh@1142 -- # return 0 00:27:42.789 18:13:35 -- spdk/autotest.sh@292 -- # run_test nvmf_dif /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/dif.sh 00:27:42.789 18:13:35 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:27:42.789 18:13:35 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:27:42.789 18:13:35 -- common/autotest_common.sh@10 -- # set +x 00:27:42.789 ************************************ 00:27:42.789 START TEST nvmf_dif 00:27:42.789 ************************************ 00:27:42.789 18:13:36 nvmf_dif -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/dif.sh 00:27:42.789 * Looking for test storage... 00:27:42.789 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:27:42.789 18:13:36 nvmf_dif -- target/dif.sh@13 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:27:42.789 18:13:36 nvmf_dif -- nvmf/common.sh@7 -- # uname -s 00:27:42.789 18:13:36 nvmf_dif -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:27:42.789 18:13:36 nvmf_dif -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:27:42.789 18:13:36 nvmf_dif -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:27:42.789 18:13:36 nvmf_dif -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:27:42.789 18:13:36 nvmf_dif -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:27:42.789 18:13:36 nvmf_dif -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:27:42.789 18:13:36 nvmf_dif -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:27:42.789 18:13:36 nvmf_dif -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:27:42.789 18:13:36 nvmf_dif -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:27:42.789 18:13:36 nvmf_dif -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:27:42.789 18:13:36 nvmf_dif -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:27:42.789 18:13:36 nvmf_dif -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:27:42.789 18:13:36 nvmf_dif -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:27:42.789 18:13:36 nvmf_dif -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:27:42.789 18:13:36 nvmf_dif -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:27:42.789 18:13:36 nvmf_dif -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:27:42.789 18:13:36 nvmf_dif -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:27:42.789 18:13:36 nvmf_dif -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:27:42.789 18:13:36 nvmf_dif -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:27:42.789 18:13:36 nvmf_dif -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:27:42.789 18:13:36 nvmf_dif -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:42.789 18:13:36 nvmf_dif -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:42.790 18:13:36 nvmf_dif -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:42.790 18:13:36 nvmf_dif -- paths/export.sh@5 -- # export PATH 00:27:42.790 18:13:36 nvmf_dif -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:27:42.790 18:13:36 nvmf_dif -- nvmf/common.sh@47 -- # : 0 00:27:42.790 18:13:36 nvmf_dif -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:27:42.790 18:13:36 nvmf_dif -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:27:42.790 18:13:36 nvmf_dif -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:27:42.790 18:13:36 nvmf_dif -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:27:42.790 18:13:36 nvmf_dif -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:27:42.790 18:13:36 nvmf_dif -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:27:42.790 18:13:36 nvmf_dif -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:27:42.790 18:13:36 nvmf_dif -- nvmf/common.sh@51 -- # have_pci_nics=0 00:27:42.790 18:13:36 nvmf_dif -- target/dif.sh@15 -- # NULL_META=16 00:27:42.790 18:13:36 nvmf_dif -- target/dif.sh@15 -- # NULL_BLOCK_SIZE=512 00:27:42.790 18:13:36 nvmf_dif -- target/dif.sh@15 -- # NULL_SIZE=64 00:27:42.790 18:13:36 nvmf_dif -- target/dif.sh@15 -- # NULL_DIF=1 00:27:42.790 18:13:36 nvmf_dif -- target/dif.sh@135 -- # nvmftestinit 00:27:42.790 18:13:36 nvmf_dif -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:27:42.790 18:13:36 nvmf_dif -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:27:42.790 18:13:36 nvmf_dif -- nvmf/common.sh@448 -- # prepare_net_devs 00:27:42.790 18:13:36 nvmf_dif -- nvmf/common.sh@410 -- # local -g is_hw=no 00:27:42.790 18:13:36 nvmf_dif -- nvmf/common.sh@412 -- # remove_spdk_ns 00:27:42.790 18:13:36 nvmf_dif -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:27:42.790 18:13:36 nvmf_dif -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:27:42.790 18:13:36 nvmf_dif -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:27:42.790 18:13:36 nvmf_dif -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:27:42.790 18:13:36 nvmf_dif -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:27:42.790 18:13:36 nvmf_dif -- nvmf/common.sh@285 -- # xtrace_disable 00:27:42.790 18:13:36 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@291 -- # pci_devs=() 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@291 -- # local -a pci_devs 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@292 -- # pci_net_devs=() 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@293 -- # pci_drivers=() 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@293 -- # local -A pci_drivers 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@295 -- # net_devs=() 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@295 -- # local -ga net_devs 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@296 -- # e810=() 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@296 -- # local -ga e810 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@297 -- # x722=() 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@297 -- # local -ga x722 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@298 -- # mlx=() 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@298 -- # local -ga mlx 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:27:48.062 Found 0000:86:00.0 (0x8086 - 0x159b) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:27:48.062 Found 0000:86:00.1 (0x8086 - 0x159b) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@390 -- # [[ up == up ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:27:48.062 Found net devices under 0000:86:00.0: cvl_0_0 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@390 -- # [[ up == up ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:27:48.062 Found net devices under 0000:86:00.1: cvl_0_1 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@414 -- # is_hw=yes 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:27:48.062 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:27:48.062 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.159 ms 00:27:48.062 00:27:48.062 --- 10.0.0.2 ping statistics --- 00:27:48.062 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:48.062 rtt min/avg/max/mdev = 0.159/0.159/0.159/0.000 ms 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:27:48.062 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:27:48.062 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.224 ms 00:27:48.062 00:27:48.062 --- 10.0.0.1 ping statistics --- 00:27:48.062 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:27:48.062 rtt min/avg/max/mdev = 0.224/0.224/0.224/0.000 ms 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@422 -- # return 0 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@450 -- # '[' iso == iso ']' 00:27:48.062 18:13:41 nvmf_dif -- nvmf/common.sh@451 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:27:49.964 0000:00:04.7 (8086 2021): Already using the vfio-pci driver 00:27:49.964 0000:5e:00.0 (8086 0a54): Already using the vfio-pci driver 00:27:49.964 0000:00:04.6 (8086 2021): Already using the vfio-pci driver 00:27:49.964 0000:00:04.5 (8086 2021): Already using the vfio-pci driver 00:27:49.964 0000:00:04.4 (8086 2021): Already using the vfio-pci driver 00:27:49.964 0000:00:04.3 (8086 2021): Already using the vfio-pci driver 00:27:49.964 0000:00:04.2 (8086 2021): Already using the vfio-pci driver 00:27:49.964 0000:00:04.1 (8086 2021): Already using the vfio-pci driver 00:27:49.964 0000:00:04.0 (8086 2021): Already using the vfio-pci driver 00:27:49.964 0000:80:04.7 (8086 2021): Already using the vfio-pci driver 00:27:49.964 0000:80:04.6 (8086 2021): Already using the vfio-pci driver 00:27:49.964 0000:80:04.5 (8086 2021): Already using the vfio-pci driver 00:27:49.964 0000:80:04.4 (8086 2021): Already using the vfio-pci driver 00:27:49.964 0000:80:04.3 (8086 2021): Already using the vfio-pci driver 00:27:49.964 0000:80:04.2 (8086 2021): Already using the vfio-pci driver 00:27:49.964 0000:80:04.1 (8086 2021): Already using the vfio-pci driver 00:27:49.964 0000:80:04.0 (8086 2021): Already using the vfio-pci driver 00:27:50.223 18:13:43 nvmf_dif -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:27:50.223 18:13:43 nvmf_dif -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:27:50.223 18:13:43 nvmf_dif -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:27:50.223 18:13:43 nvmf_dif -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:27:50.223 18:13:43 nvmf_dif -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:27:50.223 18:13:43 nvmf_dif -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:27:50.223 18:13:43 nvmf_dif -- target/dif.sh@136 -- # NVMF_TRANSPORT_OPTS+=' --dif-insert-or-strip' 00:27:50.223 18:13:43 nvmf_dif -- target/dif.sh@137 -- # nvmfappstart 00:27:50.223 18:13:43 nvmf_dif -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:27:50.223 18:13:43 nvmf_dif -- common/autotest_common.sh@722 -- # xtrace_disable 00:27:50.223 18:13:43 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:27:50.223 18:13:43 nvmf_dif -- nvmf/common.sh@481 -- # nvmfpid=760791 00:27:50.223 18:13:43 nvmf_dif -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF 00:27:50.223 18:13:43 nvmf_dif -- nvmf/common.sh@482 -- # waitforlisten 760791 00:27:50.223 18:13:43 nvmf_dif -- common/autotest_common.sh@829 -- # '[' -z 760791 ']' 00:27:50.223 18:13:43 nvmf_dif -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:27:50.223 18:13:43 nvmf_dif -- common/autotest_common.sh@834 -- # local max_retries=100 00:27:50.223 18:13:43 nvmf_dif -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:27:50.223 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:27:50.223 18:13:43 nvmf_dif -- common/autotest_common.sh@838 -- # xtrace_disable 00:27:50.223 18:13:43 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:27:50.223 [2024-07-15 18:13:43.851878] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:27:50.223 [2024-07-15 18:13:43.851920] [ DPDK EAL parameters: nvmf -c 0x1 --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:27:50.223 EAL: No free 2048 kB hugepages reported on node 1 00:27:50.223 [2024-07-15 18:13:43.912177] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:27:50.481 [2024-07-15 18:13:43.989737] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:27:50.481 [2024-07-15 18:13:43.989777] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:27:50.481 [2024-07-15 18:13:43.989785] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:27:50.481 [2024-07-15 18:13:43.989792] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:27:50.481 [2024-07-15 18:13:43.989798] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:27:50.481 [2024-07-15 18:13:43.989818] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:27:51.055 18:13:44 nvmf_dif -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:27:51.055 18:13:44 nvmf_dif -- common/autotest_common.sh@862 -- # return 0 00:27:51.055 18:13:44 nvmf_dif -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:27:51.055 18:13:44 nvmf_dif -- common/autotest_common.sh@728 -- # xtrace_disable 00:27:51.055 18:13:44 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:27:51.055 18:13:44 nvmf_dif -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:27:51.055 18:13:44 nvmf_dif -- target/dif.sh@139 -- # create_transport 00:27:51.055 18:13:44 nvmf_dif -- target/dif.sh@50 -- # rpc_cmd nvmf_create_transport -t tcp -o --dif-insert-or-strip 00:27:51.055 18:13:44 nvmf_dif -- common/autotest_common.sh@559 -- # xtrace_disable 00:27:51.055 18:13:44 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:27:51.055 [2024-07-15 18:13:44.688914] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:27:51.055 18:13:44 nvmf_dif -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:27:51.055 18:13:44 nvmf_dif -- target/dif.sh@141 -- # run_test fio_dif_1_default fio_dif_1 00:27:51.055 18:13:44 nvmf_dif -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:27:51.055 18:13:44 nvmf_dif -- common/autotest_common.sh@1105 -- # xtrace_disable 00:27:51.055 18:13:44 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:27:51.055 ************************************ 00:27:51.055 START TEST fio_dif_1_default 00:27:51.055 ************************************ 00:27:51.055 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1123 -- # fio_dif_1 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- target/dif.sh@86 -- # create_subsystems 0 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- target/dif.sh@28 -- # local sub 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- target/dif.sh@30 -- # for sub in "$@" 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- target/dif.sh@31 -- # create_subsystem 0 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- target/dif.sh@18 -- # local sub_id=0 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 1 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@559 -- # xtrace_disable 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:27:51.056 bdev_null0 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@559 -- # xtrace_disable 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@559 -- # xtrace_disable 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@559 -- # xtrace_disable 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:27:51.056 [2024-07-15 18:13:44.769332] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- target/dif.sh@87 -- # fio /dev/fd/62 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- target/dif.sh@87 -- # create_json_sub_conf 0 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- target/dif.sh@51 -- # gen_nvmf_target_json 0 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@532 -- # config=() 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@532 -- # local subsystem config 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- target/dif.sh@82 -- # gen_fio_conf 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:27:51.056 { 00:27:51.056 "params": { 00:27:51.056 "name": "Nvme$subsystem", 00:27:51.056 "trtype": "$TEST_TRANSPORT", 00:27:51.056 "traddr": "$NVMF_FIRST_TARGET_IP", 00:27:51.056 "adrfam": "ipv4", 00:27:51.056 "trsvcid": "$NVMF_PORT", 00:27:51.056 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:27:51.056 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:27:51.056 "hdgst": ${hdgst:-false}, 00:27:51.056 "ddgst": ${ddgst:-false} 00:27:51.056 }, 00:27:51.056 "method": "bdev_nvme_attach_controller" 00:27:51.056 } 00:27:51.056 EOF 00:27:51.056 )") 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- target/dif.sh@54 -- # local file 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- target/dif.sh@56 -- # cat 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1339 -- # local sanitizers 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1341 -- # shift 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1343 -- # local asan_lib= 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@554 -- # cat 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- target/dif.sh@72 -- # (( file = 1 )) 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- target/dif.sh@72 -- # (( file <= files )) 00:27:51.056 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # grep libasan 00:27:51.315 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:27:51.315 18:13:44 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@556 -- # jq . 00:27:51.315 18:13:44 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@557 -- # IFS=, 00:27:51.315 18:13:44 nvmf_dif.fio_dif_1_default -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:27:51.315 "params": { 00:27:51.315 "name": "Nvme0", 00:27:51.315 "trtype": "tcp", 00:27:51.315 "traddr": "10.0.0.2", 00:27:51.315 "adrfam": "ipv4", 00:27:51.315 "trsvcid": "4420", 00:27:51.315 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:27:51.315 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:27:51.315 "hdgst": false, 00:27:51.315 "ddgst": false 00:27:51.315 }, 00:27:51.315 "method": "bdev_nvme_attach_controller" 00:27:51.315 }' 00:27:51.315 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # asan_lib= 00:27:51.315 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:27:51.315 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:27:51.315 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:27:51.315 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:27:51.315 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:27:51.315 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1345 -- # asan_lib= 00:27:51.315 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:27:51.315 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:27:51.315 18:13:44 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:27:51.585 filename0: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=4 00:27:51.585 fio-3.35 00:27:51.585 Starting 1 thread 00:27:51.585 EAL: No free 2048 kB hugepages reported on node 1 00:28:03.839 00:28:03.839 filename0: (groupid=0, jobs=1): err= 0: pid=761249: Mon Jul 15 18:13:55 2024 00:28:03.839 read: IOPS=96, BW=385KiB/s (394kB/s)(3856KiB/10019msec) 00:28:03.839 slat (nsec): min=5906, max=63991, avg=6696.25, stdev=2468.88 00:28:03.839 clat (usec): min=40844, max=44518, avg=41551.77, stdev=522.30 00:28:03.839 lat (usec): min=40850, max=44549, avg=41558.47, stdev=522.53 00:28:03.839 clat percentiles (usec): 00:28:03.839 | 1.00th=[40633], 5.00th=[41157], 10.00th=[41157], 20.00th=[41157], 00:28:03.839 | 30.00th=[41157], 40.00th=[41157], 50.00th=[41681], 60.00th=[42206], 00:28:03.839 | 70.00th=[42206], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:28:03.839 | 99.00th=[42206], 99.50th=[42206], 99.90th=[44303], 99.95th=[44303], 00:28:03.839 | 99.99th=[44303] 00:28:03.839 bw ( KiB/s): min= 352, max= 416, per=99.77%, avg=384.00, stdev=10.38, samples=20 00:28:03.839 iops : min= 88, max= 104, avg=96.00, stdev= 2.60, samples=20 00:28:03.839 lat (msec) : 50=100.00% 00:28:03.839 cpu : usr=94.31%, sys=5.38%, ctx=42, majf=0, minf=211 00:28:03.839 IO depths : 1=25.0%, 2=50.0%, 4=25.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:28:03.839 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:03.839 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:03.839 issued rwts: total=964,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:03.839 latency : target=0, window=0, percentile=100.00%, depth=4 00:28:03.839 00:28:03.839 Run status group 0 (all jobs): 00:28:03.839 READ: bw=385KiB/s (394kB/s), 385KiB/s-385KiB/s (394kB/s-394kB/s), io=3856KiB (3949kB), run=10019-10019msec 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_default -- target/dif.sh@88 -- # destroy_subsystems 0 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_default -- target/dif.sh@43 -- # local sub 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_default -- target/dif.sh@45 -- # for sub in "$@" 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_default -- target/dif.sh@46 -- # destroy_subsystem 0 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_default -- target/dif.sh@36 -- # local sub_id=0 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_default -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_default -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:03.839 00:28:03.839 real 0m11.045s 00:28:03.839 user 0m16.297s 00:28:03.839 sys 0m0.810s 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@1124 -- # xtrace_disable 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_default -- common/autotest_common.sh@10 -- # set +x 00:28:03.839 ************************************ 00:28:03.839 END TEST fio_dif_1_default 00:28:03.839 ************************************ 00:28:03.839 18:13:55 nvmf_dif -- common/autotest_common.sh@1142 -- # return 0 00:28:03.839 18:13:55 nvmf_dif -- target/dif.sh@142 -- # run_test fio_dif_1_multi_subsystems fio_dif_1_multi_subsystems 00:28:03.839 18:13:55 nvmf_dif -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:28:03.839 18:13:55 nvmf_dif -- common/autotest_common.sh@1105 -- # xtrace_disable 00:28:03.839 18:13:55 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:28:03.839 ************************************ 00:28:03.839 START TEST fio_dif_1_multi_subsystems 00:28:03.839 ************************************ 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1123 -- # fio_dif_1_multi_subsystems 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@92 -- # local files=1 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@94 -- # create_subsystems 0 1 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@28 -- # local sub 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@30 -- # for sub in "$@" 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@31 -- # create_subsystem 0 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@18 -- # local sub_id=0 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 1 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:03.839 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:28:03.840 bdev_null0 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:28:03.840 [2024-07-15 18:13:55.878896] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@30 -- # for sub in "$@" 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@31 -- # create_subsystem 1 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@18 -- # local sub_id=1 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null1 64 512 --md-size 16 --dif-type 1 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:28:03.840 bdev_null1 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 --serial-number 53313233-1 --allow-any-host 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 bdev_null1 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@95 -- # fio /dev/fd/62 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@95 -- # create_json_sub_conf 0 1 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@51 -- # gen_nvmf_target_json 0 1 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@532 -- # config=() 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@532 -- # local subsystem config 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@82 -- # gen_fio_conf 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:28:03.840 { 00:28:03.840 "params": { 00:28:03.840 "name": "Nvme$subsystem", 00:28:03.840 "trtype": "$TEST_TRANSPORT", 00:28:03.840 "traddr": "$NVMF_FIRST_TARGET_IP", 00:28:03.840 "adrfam": "ipv4", 00:28:03.840 "trsvcid": "$NVMF_PORT", 00:28:03.840 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:28:03.840 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:28:03.840 "hdgst": ${hdgst:-false}, 00:28:03.840 "ddgst": ${ddgst:-false} 00:28:03.840 }, 00:28:03.840 "method": "bdev_nvme_attach_controller" 00:28:03.840 } 00:28:03.840 EOF 00:28:03.840 )") 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@54 -- # local file 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@56 -- # cat 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1339 -- # local sanitizers 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1341 -- # shift 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1343 -- # local asan_lib= 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@554 -- # cat 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@72 -- # (( file = 1 )) 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@72 -- # (( file <= files )) 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@73 -- # cat 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # grep libasan 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:28:03.840 { 00:28:03.840 "params": { 00:28:03.840 "name": "Nvme$subsystem", 00:28:03.840 "trtype": "$TEST_TRANSPORT", 00:28:03.840 "traddr": "$NVMF_FIRST_TARGET_IP", 00:28:03.840 "adrfam": "ipv4", 00:28:03.840 "trsvcid": "$NVMF_PORT", 00:28:03.840 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:28:03.840 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:28:03.840 "hdgst": ${hdgst:-false}, 00:28:03.840 "ddgst": ${ddgst:-false} 00:28:03.840 }, 00:28:03.840 "method": "bdev_nvme_attach_controller" 00:28:03.840 } 00:28:03.840 EOF 00:28:03.840 )") 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@72 -- # (( file++ )) 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@72 -- # (( file <= files )) 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@554 -- # cat 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@556 -- # jq . 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@557 -- # IFS=, 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:28:03.840 "params": { 00:28:03.840 "name": "Nvme0", 00:28:03.840 "trtype": "tcp", 00:28:03.840 "traddr": "10.0.0.2", 00:28:03.840 "adrfam": "ipv4", 00:28:03.840 "trsvcid": "4420", 00:28:03.840 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:28:03.840 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:28:03.840 "hdgst": false, 00:28:03.840 "ddgst": false 00:28:03.840 }, 00:28:03.840 "method": "bdev_nvme_attach_controller" 00:28:03.840 },{ 00:28:03.840 "params": { 00:28:03.840 "name": "Nvme1", 00:28:03.840 "trtype": "tcp", 00:28:03.840 "traddr": "10.0.0.2", 00:28:03.840 "adrfam": "ipv4", 00:28:03.840 "trsvcid": "4420", 00:28:03.840 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:28:03.840 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:28:03.840 "hdgst": false, 00:28:03.840 "ddgst": false 00:28:03.840 }, 00:28:03.840 "method": "bdev_nvme_attach_controller" 00:28:03.840 }' 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # asan_lib= 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1345 -- # asan_lib= 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:28:03.840 18:13:55 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:28:03.840 filename0: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=4 00:28:03.840 filename1: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=4 00:28:03.840 fio-3.35 00:28:03.840 Starting 2 threads 00:28:03.840 EAL: No free 2048 kB hugepages reported on node 1 00:28:13.837 00:28:13.837 filename0: (groupid=0, jobs=1): err= 0: pid=763287: Mon Jul 15 18:14:07 2024 00:28:13.837 read: IOPS=140, BW=560KiB/s (574kB/s)(5616KiB/10020msec) 00:28:13.837 slat (nsec): min=6177, max=46494, avg=7573.76, stdev=2700.09 00:28:13.837 clat (usec): min=541, max=42724, avg=28522.88, stdev=19116.93 00:28:13.837 lat (usec): min=548, max=42737, avg=28530.45, stdev=19116.69 00:28:13.837 clat percentiles (usec): 00:28:13.837 | 1.00th=[ 553], 5.00th=[ 570], 10.00th=[ 570], 20.00th=[ 594], 00:28:13.837 | 30.00th=[ 693], 40.00th=[41157], 50.00th=[41157], 60.00th=[41157], 00:28:13.837 | 70.00th=[41681], 80.00th=[42206], 90.00th=[42206], 95.00th=[42206], 00:28:13.837 | 99.00th=[42206], 99.50th=[42730], 99.90th=[42730], 99.95th=[42730], 00:28:13.837 | 99.99th=[42730] 00:28:13.837 bw ( KiB/s): min= 384, max= 768, per=42.51%, avg=560.00, stdev=175.12, samples=20 00:28:13.837 iops : min= 96, max= 192, avg=140.00, stdev=43.78, samples=20 00:28:13.837 lat (usec) : 750=31.34%, 1000=0.28% 00:28:13.837 lat (msec) : 2=0.28%, 50=68.09% 00:28:13.837 cpu : usr=97.64%, sys=2.07%, ctx=14, majf=0, minf=108 00:28:13.837 IO depths : 1=25.0%, 2=50.0%, 4=25.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:28:13.837 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:13.837 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:13.837 issued rwts: total=1404,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:13.837 latency : target=0, window=0, percentile=100.00%, depth=4 00:28:13.837 filename1: (groupid=0, jobs=1): err= 0: pid=763288: Mon Jul 15 18:14:07 2024 00:28:13.837 read: IOPS=189, BW=757KiB/s (775kB/s)(7584KiB/10020msec) 00:28:13.837 slat (nsec): min=6151, max=59238, avg=7296.19, stdev=2616.45 00:28:13.837 clat (usec): min=543, max=42521, avg=21116.57, stdev=20424.88 00:28:13.837 lat (usec): min=549, max=42528, avg=21123.86, stdev=20424.12 00:28:13.837 clat percentiles (usec): 00:28:13.837 | 1.00th=[ 553], 5.00th=[ 570], 10.00th=[ 578], 20.00th=[ 594], 00:28:13.837 | 30.00th=[ 603], 40.00th=[ 619], 50.00th=[41157], 60.00th=[41157], 00:28:13.837 | 70.00th=[41157], 80.00th=[41157], 90.00th=[41681], 95.00th=[42206], 00:28:13.837 | 99.00th=[42206], 99.50th=[42206], 99.90th=[42730], 99.95th=[42730], 00:28:13.837 | 99.99th=[42730] 00:28:13.837 bw ( KiB/s): min= 672, max= 768, per=57.39%, avg=756.80, stdev=28.00, samples=20 00:28:13.837 iops : min= 168, max= 192, avg=189.20, stdev= 7.00, samples=20 00:28:13.837 lat (usec) : 750=46.10%, 1000=3.69% 00:28:13.837 lat (msec) : 50=50.21% 00:28:13.837 cpu : usr=97.89%, sys=1.82%, ctx=15, majf=0, minf=196 00:28:13.837 IO depths : 1=25.0%, 2=50.0%, 4=25.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:28:13.837 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:13.837 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:13.837 issued rwts: total=1896,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:13.837 latency : target=0, window=0, percentile=100.00%, depth=4 00:28:13.837 00:28:13.837 Run status group 0 (all jobs): 00:28:13.837 READ: bw=1317KiB/s (1349kB/s), 560KiB/s-757KiB/s (574kB/s-775kB/s), io=12.9MiB (13.5MB), run=10020-10020msec 00:28:13.837 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@96 -- # destroy_subsystems 0 1 00:28:13.837 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@43 -- # local sub 00:28:13.837 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@45 -- # for sub in "$@" 00:28:13.837 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@46 -- # destroy_subsystem 0 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@36 -- # local sub_id=0 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@45 -- # for sub in "$@" 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@46 -- # destroy_subsystem 1 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@36 -- # local sub_id=1 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null1 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:13.838 00:28:13.838 real 0m11.423s 00:28:13.838 user 0m26.023s 00:28:13.838 sys 0m0.696s 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@1124 -- # xtrace_disable 00:28:13.838 18:14:07 nvmf_dif.fio_dif_1_multi_subsystems -- common/autotest_common.sh@10 -- # set +x 00:28:13.838 ************************************ 00:28:13.838 END TEST fio_dif_1_multi_subsystems 00:28:13.838 ************************************ 00:28:13.838 18:14:07 nvmf_dif -- common/autotest_common.sh@1142 -- # return 0 00:28:13.838 18:14:07 nvmf_dif -- target/dif.sh@143 -- # run_test fio_dif_rand_params fio_dif_rand_params 00:28:13.838 18:14:07 nvmf_dif -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:28:13.838 18:14:07 nvmf_dif -- common/autotest_common.sh@1105 -- # xtrace_disable 00:28:13.838 18:14:07 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:28:13.838 ************************************ 00:28:13.838 START TEST fio_dif_rand_params 00:28:13.838 ************************************ 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1123 -- # fio_dif_rand_params 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@100 -- # local NULL_DIF 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@101 -- # local bs numjobs runtime iodepth files 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # NULL_DIF=3 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # bs=128k 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # numjobs=3 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # iodepth=3 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@103 -- # runtime=5 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@105 -- # create_subsystems 0 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@28 -- # local sub 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 0 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=0 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 3 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:13.838 bdev_null0 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:13.838 [2024-07-15 18:14:07.368367] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@106 -- # fio /dev/fd/62 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@106 -- # create_json_sub_conf 0 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@51 -- # gen_nvmf_target_json 0 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@532 -- # config=() 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@532 -- # local subsystem config 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # gen_fio_conf 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:28:13.838 { 00:28:13.838 "params": { 00:28:13.838 "name": "Nvme$subsystem", 00:28:13.838 "trtype": "$TEST_TRANSPORT", 00:28:13.838 "traddr": "$NVMF_FIRST_TARGET_IP", 00:28:13.838 "adrfam": "ipv4", 00:28:13.838 "trsvcid": "$NVMF_PORT", 00:28:13.838 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:28:13.838 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:28:13.838 "hdgst": ${hdgst:-false}, 00:28:13.838 "ddgst": ${ddgst:-false} 00:28:13.838 }, 00:28:13.838 "method": "bdev_nvme_attach_controller" 00:28:13.838 } 00:28:13.838 EOF 00:28:13.838 )") 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@54 -- # local file 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@56 -- # cat 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1339 -- # local sanitizers 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1341 -- # shift 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # local asan_lib= 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@554 -- # cat 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file = 1 )) 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # grep libasan 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@556 -- # jq . 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@557 -- # IFS=, 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:28:13.838 "params": { 00:28:13.838 "name": "Nvme0", 00:28:13.838 "trtype": "tcp", 00:28:13.838 "traddr": "10.0.0.2", 00:28:13.838 "adrfam": "ipv4", 00:28:13.838 "trsvcid": "4420", 00:28:13.838 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:28:13.838 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:28:13.838 "hdgst": false, 00:28:13.838 "ddgst": false 00:28:13.838 }, 00:28:13.838 "method": "bdev_nvme_attach_controller" 00:28:13.838 }' 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # asan_lib= 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # asan_lib= 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:28:13.838 18:14:07 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:28:14.097 filename0: (g=0): rw=randread, bs=(R) 128KiB-128KiB, (W) 128KiB-128KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=3 00:28:14.097 ... 00:28:14.097 fio-3.35 00:28:14.097 Starting 3 threads 00:28:14.097 EAL: No free 2048 kB hugepages reported on node 1 00:28:20.669 00:28:20.669 filename0: (groupid=0, jobs=1): err= 0: pid=765294: Mon Jul 15 18:14:13 2024 00:28:20.669 read: IOPS=275, BW=34.4MiB/s (36.1MB/s)(172MiB/5007msec) 00:28:20.669 slat (nsec): min=6221, max=36778, avg=9386.21, stdev=2848.59 00:28:20.669 clat (usec): min=3766, max=92029, avg=10883.96, stdev=12766.32 00:28:20.669 lat (usec): min=3772, max=92043, avg=10893.34, stdev=12766.54 00:28:20.669 clat percentiles (usec): 00:28:20.669 | 1.00th=[ 4178], 5.00th=[ 4424], 10.00th=[ 4686], 20.00th=[ 5276], 00:28:20.669 | 30.00th=[ 5997], 40.00th=[ 6456], 50.00th=[ 6718], 60.00th=[ 7177], 00:28:20.669 | 70.00th=[ 8160], 80.00th=[ 8979], 90.00th=[10945], 95.00th=[47973], 00:28:20.669 | 99.00th=[50070], 99.50th=[50594], 99.90th=[87557], 99.95th=[91751], 00:28:20.669 | 99.99th=[91751] 00:28:20.669 bw ( KiB/s): min=22272, max=42752, per=35.09%, avg=35200.00, stdev=6715.09, samples=10 00:28:20.669 iops : min= 174, max= 334, avg=275.00, stdev=52.46, samples=10 00:28:20.669 lat (msec) : 4=0.07%, 10=88.75%, 20=1.31%, 50=8.42%, 100=1.45% 00:28:20.669 cpu : usr=94.47%, sys=5.17%, ctx=5, majf=0, minf=117 00:28:20.669 IO depths : 1=1.7%, 2=98.3%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:28:20.669 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:20.669 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:20.669 issued rwts: total=1378,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:20.669 latency : target=0, window=0, percentile=100.00%, depth=3 00:28:20.669 filename0: (groupid=0, jobs=1): err= 0: pid=765295: Mon Jul 15 18:14:13 2024 00:28:20.669 read: IOPS=237, BW=29.7MiB/s (31.2MB/s)(150MiB/5043msec) 00:28:20.669 slat (usec): min=6, max=133, avg= 9.49, stdev= 4.38 00:28:20.669 clat (usec): min=3678, max=91364, avg=12571.90, stdev=14462.09 00:28:20.669 lat (usec): min=3685, max=91376, avg=12581.40, stdev=14462.35 00:28:20.669 clat percentiles (usec): 00:28:20.669 | 1.00th=[ 4080], 5.00th=[ 4359], 10.00th=[ 4555], 20.00th=[ 5014], 00:28:20.669 | 30.00th=[ 6194], 40.00th=[ 6652], 50.00th=[ 7046], 60.00th=[ 7701], 00:28:20.669 | 70.00th=[ 9110], 80.00th=[10028], 90.00th=[47449], 95.00th=[49021], 00:28:20.669 | 99.00th=[51643], 99.50th=[51643], 99.90th=[53740], 99.95th=[91751], 00:28:20.669 | 99.99th=[91751] 00:28:20.669 bw ( KiB/s): min=23808, max=43520, per=30.52%, avg=30617.60, stdev=5908.62, samples=10 00:28:20.669 iops : min= 186, max= 340, avg=239.20, stdev=46.16, samples=10 00:28:20.669 lat (msec) : 4=0.67%, 10=79.32%, 20=6.67%, 50=10.43%, 100=2.92% 00:28:20.669 cpu : usr=95.54%, sys=4.13%, ctx=7, majf=0, minf=30 00:28:20.669 IO depths : 1=0.3%, 2=99.7%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:28:20.669 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:20.669 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:20.669 issued rwts: total=1199,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:20.669 latency : target=0, window=0, percentile=100.00%, depth=3 00:28:20.669 filename0: (groupid=0, jobs=1): err= 0: pid=765296: Mon Jul 15 18:14:13 2024 00:28:20.669 read: IOPS=274, BW=34.4MiB/s (36.0MB/s)(172MiB/5003msec) 00:28:20.669 slat (nsec): min=6222, max=97734, avg=9691.53, stdev=3509.61 00:28:20.669 clat (usec): min=3919, max=89978, avg=10901.25, stdev=12510.74 00:28:20.669 lat (usec): min=3926, max=89989, avg=10910.94, stdev=12511.17 00:28:20.669 clat percentiles (usec): 00:28:20.669 | 1.00th=[ 4080], 5.00th=[ 4359], 10.00th=[ 4621], 20.00th=[ 5276], 00:28:20.669 | 30.00th=[ 6259], 40.00th=[ 6718], 50.00th=[ 6980], 60.00th=[ 7439], 00:28:20.669 | 70.00th=[ 8455], 80.00th=[ 9503], 90.00th=[11863], 95.00th=[47973], 00:28:20.669 | 99.00th=[50594], 99.50th=[51119], 99.90th=[88605], 99.95th=[89654], 00:28:20.669 | 99.99th=[89654] 00:28:20.669 bw ( KiB/s): min=29440, max=44633, per=35.02%, avg=35132.10, stdev=5281.92, samples=10 00:28:20.669 iops : min= 230, max= 348, avg=274.40, stdev=41.13, samples=10 00:28:20.669 lat (msec) : 4=0.22%, 10=84.00%, 20=6.62%, 50=7.42%, 100=1.75% 00:28:20.669 cpu : usr=94.84%, sys=4.86%, ctx=9, majf=0, minf=118 00:28:20.669 IO depths : 1=0.5%, 2=99.5%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:28:20.669 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:20.669 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:20.669 issued rwts: total=1375,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:20.669 latency : target=0, window=0, percentile=100.00%, depth=3 00:28:20.669 00:28:20.669 Run status group 0 (all jobs): 00:28:20.669 READ: bw=98.0MiB/s (103MB/s), 29.7MiB/s-34.4MiB/s (31.2MB/s-36.1MB/s), io=494MiB (518MB), run=5003-5043msec 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@107 -- # destroy_subsystems 0 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@43 -- # local sub 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 0 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=0 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # NULL_DIF=2 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # bs=4k 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # numjobs=8 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # iodepth=16 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # runtime= 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@109 -- # files=2 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@111 -- # create_subsystems 0 1 2 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@28 -- # local sub 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 0 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=0 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 2 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:20.669 bdev_null0 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:20.669 [2024-07-15 18:14:13.543723] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 1 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=1 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null1 64 512 --md-size 16 --dif-type 2 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:20.669 bdev_null1 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 --serial-number 53313233-1 --allow-any-host 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 bdev_null1 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 2 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=2 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null2 64 512 --md-size 16 --dif-type 2 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:20.669 bdev_null2 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:20.669 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode2 --serial-number 53313233-2 --allow-any-host 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode2 bdev_null2 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode2 -t tcp -a 10.0.0.2 -s 4420 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@112 -- # fio /dev/fd/62 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@112 -- # create_json_sub_conf 0 1 2 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@51 -- # gen_nvmf_target_json 0 1 2 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@532 -- # config=() 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@532 -- # local subsystem config 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # gen_fio_conf 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:28:20.670 { 00:28:20.670 "params": { 00:28:20.670 "name": "Nvme$subsystem", 00:28:20.670 "trtype": "$TEST_TRANSPORT", 00:28:20.670 "traddr": "$NVMF_FIRST_TARGET_IP", 00:28:20.670 "adrfam": "ipv4", 00:28:20.670 "trsvcid": "$NVMF_PORT", 00:28:20.670 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:28:20.670 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:28:20.670 "hdgst": ${hdgst:-false}, 00:28:20.670 "ddgst": ${ddgst:-false} 00:28:20.670 }, 00:28:20.670 "method": "bdev_nvme_attach_controller" 00:28:20.670 } 00:28:20.670 EOF 00:28:20.670 )") 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@54 -- # local file 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@56 -- # cat 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1339 -- # local sanitizers 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1341 -- # shift 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # local asan_lib= 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@554 -- # cat 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file = 1 )) 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # grep libasan 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@73 -- # cat 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:28:20.670 { 00:28:20.670 "params": { 00:28:20.670 "name": "Nvme$subsystem", 00:28:20.670 "trtype": "$TEST_TRANSPORT", 00:28:20.670 "traddr": "$NVMF_FIRST_TARGET_IP", 00:28:20.670 "adrfam": "ipv4", 00:28:20.670 "trsvcid": "$NVMF_PORT", 00:28:20.670 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:28:20.670 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:28:20.670 "hdgst": ${hdgst:-false}, 00:28:20.670 "ddgst": ${ddgst:-false} 00:28:20.670 }, 00:28:20.670 "method": "bdev_nvme_attach_controller" 00:28:20.670 } 00:28:20.670 EOF 00:28:20.670 )") 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@554 -- # cat 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file++ )) 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@73 -- # cat 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file++ )) 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:28:20.670 { 00:28:20.670 "params": { 00:28:20.670 "name": "Nvme$subsystem", 00:28:20.670 "trtype": "$TEST_TRANSPORT", 00:28:20.670 "traddr": "$NVMF_FIRST_TARGET_IP", 00:28:20.670 "adrfam": "ipv4", 00:28:20.670 "trsvcid": "$NVMF_PORT", 00:28:20.670 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:28:20.670 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:28:20.670 "hdgst": ${hdgst:-false}, 00:28:20.670 "ddgst": ${ddgst:-false} 00:28:20.670 }, 00:28:20.670 "method": "bdev_nvme_attach_controller" 00:28:20.670 } 00:28:20.670 EOF 00:28:20.670 )") 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@554 -- # cat 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@556 -- # jq . 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@557 -- # IFS=, 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:28:20.670 "params": { 00:28:20.670 "name": "Nvme0", 00:28:20.670 "trtype": "tcp", 00:28:20.670 "traddr": "10.0.0.2", 00:28:20.670 "adrfam": "ipv4", 00:28:20.670 "trsvcid": "4420", 00:28:20.670 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:28:20.670 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:28:20.670 "hdgst": false, 00:28:20.670 "ddgst": false 00:28:20.670 }, 00:28:20.670 "method": "bdev_nvme_attach_controller" 00:28:20.670 },{ 00:28:20.670 "params": { 00:28:20.670 "name": "Nvme1", 00:28:20.670 "trtype": "tcp", 00:28:20.670 "traddr": "10.0.0.2", 00:28:20.670 "adrfam": "ipv4", 00:28:20.670 "trsvcid": "4420", 00:28:20.670 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:28:20.670 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:28:20.670 "hdgst": false, 00:28:20.670 "ddgst": false 00:28:20.670 }, 00:28:20.670 "method": "bdev_nvme_attach_controller" 00:28:20.670 },{ 00:28:20.670 "params": { 00:28:20.670 "name": "Nvme2", 00:28:20.670 "trtype": "tcp", 00:28:20.670 "traddr": "10.0.0.2", 00:28:20.670 "adrfam": "ipv4", 00:28:20.670 "trsvcid": "4420", 00:28:20.670 "subnqn": "nqn.2016-06.io.spdk:cnode2", 00:28:20.670 "hostnqn": "nqn.2016-06.io.spdk:host2", 00:28:20.670 "hdgst": false, 00:28:20.670 "ddgst": false 00:28:20.670 }, 00:28:20.670 "method": "bdev_nvme_attach_controller" 00:28:20.670 }' 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # asan_lib= 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # asan_lib= 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:28:20.670 18:14:13 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:28:20.670 filename0: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=16 00:28:20.670 ... 00:28:20.670 filename1: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=16 00:28:20.670 ... 00:28:20.670 filename2: (g=0): rw=randread, bs=(R) 4096B-4096B, (W) 4096B-4096B, (T) 4096B-4096B, ioengine=spdk_bdev, iodepth=16 00:28:20.670 ... 00:28:20.670 fio-3.35 00:28:20.670 Starting 24 threads 00:28:20.670 EAL: No free 2048 kB hugepages reported on node 1 00:28:32.867 00:28:32.867 filename0: (groupid=0, jobs=1): err= 0: pid=766368: Mon Jul 15 18:14:25 2024 00:28:32.867 read: IOPS=572, BW=2288KiB/s (2343kB/s)(22.4MiB/10012msec) 00:28:32.867 slat (nsec): min=7066, max=97295, avg=45804.10, stdev=19967.26 00:28:32.867 clat (usec): min=19069, max=37119, avg=27499.38, stdev=806.36 00:28:32.867 lat (usec): min=19085, max=37153, avg=27545.19, stdev=809.18 00:28:32.867 clat percentiles (usec): 00:28:32.867 | 1.00th=[26608], 5.00th=[27132], 10.00th=[27132], 20.00th=[27132], 00:28:32.867 | 30.00th=[27132], 40.00th=[27395], 50.00th=[27395], 60.00th=[27657], 00:28:32.867 | 70.00th=[27657], 80.00th=[27657], 90.00th=[27919], 95.00th=[28443], 00:28:32.867 | 99.00th=[28967], 99.50th=[29230], 99.90th=[36963], 99.95th=[36963], 00:28:32.867 | 99.99th=[36963] 00:28:32.867 bw ( KiB/s): min= 2176, max= 2304, per=4.14%, avg=2288.95, stdev=39.87, samples=19 00:28:32.867 iops : min= 544, max= 576, avg=572.00, stdev= 9.91, samples=19 00:28:32.867 lat (msec) : 20=0.28%, 50=99.72% 00:28:32.867 cpu : usr=98.73%, sys=0.86%, ctx=9, majf=0, minf=9 00:28:32.867 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.2%, 32=0.0%, >=64=0.0% 00:28:32.867 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.867 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.867 issued rwts: total=5728,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.867 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.867 filename0: (groupid=0, jobs=1): err= 0: pid=766369: Mon Jul 15 18:14:25 2024 00:28:32.867 read: IOPS=593, BW=2372KiB/s (2429kB/s)(23.2MiB/10005msec) 00:28:32.867 slat (nsec): min=6795, max=90833, avg=21485.99, stdev=15378.82 00:28:32.867 clat (usec): min=6140, max=46617, avg=26790.09, stdev=3612.64 00:28:32.867 lat (usec): min=6153, max=46631, avg=26811.57, stdev=3614.60 00:28:32.867 clat percentiles (usec): 00:28:32.867 | 1.00th=[17171], 5.00th=[19006], 10.00th=[22152], 20.00th=[27132], 00:28:32.867 | 30.00th=[27395], 40.00th=[27395], 50.00th=[27657], 60.00th=[27657], 00:28:32.867 | 70.00th=[27657], 80.00th=[27919], 90.00th=[28443], 95.00th=[31589], 00:28:32.867 | 99.00th=[37487], 99.50th=[39060], 99.90th=[46400], 99.95th=[46400], 00:28:32.867 | 99.99th=[46400] 00:28:32.867 bw ( KiB/s): min= 2128, max= 2848, per=4.28%, avg=2362.68, stdev=159.20, samples=19 00:28:32.867 iops : min= 532, max= 712, avg=590.47, stdev=39.88, samples=19 00:28:32.867 lat (msec) : 10=0.27%, 20=7.33%, 50=92.40% 00:28:32.867 cpu : usr=99.00%, sys=0.62%, ctx=17, majf=0, minf=9 00:28:32.867 IO depths : 1=3.6%, 2=7.2%, 4=15.7%, 8=63.4%, 16=10.1%, 32=0.0%, >=64=0.0% 00:28:32.867 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.867 complete : 0=0.0%, 4=91.7%, 8=3.8%, 16=4.5%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.867 issued rwts: total=5934,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.867 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.867 filename0: (groupid=0, jobs=1): err= 0: pid=766370: Mon Jul 15 18:14:25 2024 00:28:32.867 read: IOPS=572, BW=2288KiB/s (2343kB/s)(22.4MiB/10013msec) 00:28:32.867 slat (nsec): min=6962, max=89947, avg=20524.72, stdev=13007.16 00:28:32.867 clat (usec): min=23727, max=37547, avg=27803.68, stdev=674.08 00:28:32.867 lat (usec): min=23756, max=37578, avg=27824.20, stdev=671.86 00:28:32.867 clat percentiles (usec): 00:28:32.867 | 1.00th=[26870], 5.00th=[27395], 10.00th=[27395], 20.00th=[27657], 00:28:32.867 | 30.00th=[27657], 40.00th=[27657], 50.00th=[27657], 60.00th=[27657], 00:28:32.867 | 70.00th=[27919], 80.00th=[27919], 90.00th=[28443], 95.00th=[28705], 00:28:32.867 | 99.00th=[29230], 99.50th=[29492], 99.90th=[37487], 99.95th=[37487], 00:28:32.867 | 99.99th=[37487] 00:28:32.867 bw ( KiB/s): min= 2176, max= 2304, per=4.13%, avg=2282.21, stdev=47.31, samples=19 00:28:32.867 iops : min= 544, max= 576, avg=570.32, stdev=11.74, samples=19 00:28:32.867 lat (msec) : 50=100.00% 00:28:32.867 cpu : usr=98.96%, sys=0.66%, ctx=10, majf=0, minf=9 00:28:32.867 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.2%, 32=0.0%, >=64=0.0% 00:28:32.867 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.867 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.867 issued rwts: total=5728,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.867 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.867 filename0: (groupid=0, jobs=1): err= 0: pid=766371: Mon Jul 15 18:14:25 2024 00:28:32.867 read: IOPS=575, BW=2303KiB/s (2358kB/s)(22.5MiB/10004msec) 00:28:32.867 slat (nsec): min=6844, max=80009, avg=37574.54, stdev=12685.24 00:28:32.867 clat (usec): min=8104, max=29414, avg=27471.69, stdev=1579.72 00:28:32.867 lat (usec): min=8111, max=29438, avg=27509.27, stdev=1581.32 00:28:32.867 clat percentiles (usec): 00:28:32.867 | 1.00th=[19268], 5.00th=[27132], 10.00th=[27132], 20.00th=[27395], 00:28:32.867 | 30.00th=[27395], 40.00th=[27395], 50.00th=[27657], 60.00th=[27657], 00:28:32.867 | 70.00th=[27657], 80.00th=[27919], 90.00th=[28181], 95.00th=[28443], 00:28:32.867 | 99.00th=[28967], 99.50th=[28967], 99.90th=[29230], 99.95th=[29230], 00:28:32.867 | 99.99th=[29492] 00:28:32.867 bw ( KiB/s): min= 2176, max= 2432, per=4.17%, avg=2302.42, stdev=42.73, samples=19 00:28:32.867 iops : min= 544, max= 608, avg=575.37, stdev=10.71, samples=19 00:28:32.867 lat (msec) : 10=0.52%, 20=0.61%, 50=98.87% 00:28:32.867 cpu : usr=99.17%, sys=0.50%, ctx=11, majf=0, minf=9 00:28:32.867 IO depths : 1=6.2%, 2=12.4%, 4=24.9%, 8=50.2%, 16=6.3%, 32=0.0%, >=64=0.0% 00:28:32.867 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.867 complete : 0=0.0%, 4=94.1%, 8=0.1%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.867 issued rwts: total=5760,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.867 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.867 filename0: (groupid=0, jobs=1): err= 0: pid=766372: Mon Jul 15 18:14:25 2024 00:28:32.867 read: IOPS=573, BW=2294KiB/s (2349kB/s)(22.4MiB/10015msec) 00:28:32.867 slat (nsec): min=7020, max=79615, avg=21265.95, stdev=10829.07 00:28:32.867 clat (usec): min=11729, max=39176, avg=27702.80, stdev=884.60 00:28:32.867 lat (usec): min=11747, max=39219, avg=27724.07, stdev=884.04 00:28:32.867 clat percentiles (usec): 00:28:32.867 | 1.00th=[26346], 5.00th=[27132], 10.00th=[27395], 20.00th=[27395], 00:28:32.867 | 30.00th=[27657], 40.00th=[27657], 50.00th=[27657], 60.00th=[27657], 00:28:32.867 | 70.00th=[27919], 80.00th=[27919], 90.00th=[28443], 95.00th=[28443], 00:28:32.867 | 99.00th=[29230], 99.50th=[30016], 99.90th=[30540], 99.95th=[36439], 00:28:32.867 | 99.99th=[39060] 00:28:32.867 bw ( KiB/s): min= 2176, max= 2304, per=4.14%, avg=2288.95, stdev=39.87, samples=19 00:28:32.867 iops : min= 544, max= 576, avg=572.00, stdev= 9.91, samples=19 00:28:32.867 lat (msec) : 20=0.31%, 50=99.69% 00:28:32.867 cpu : usr=98.61%, sys=1.00%, ctx=17, majf=0, minf=9 00:28:32.867 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.3%, 32=0.0%, >=64=0.0% 00:28:32.867 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.867 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.867 issued rwts: total=5744,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.867 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.867 filename0: (groupid=0, jobs=1): err= 0: pid=766373: Mon Jul 15 18:14:25 2024 00:28:32.867 read: IOPS=572, BW=2291KiB/s (2346kB/s)(22.4MiB/10002msec) 00:28:32.867 slat (nsec): min=6125, max=76262, avg=21711.60, stdev=9538.96 00:28:32.867 clat (usec): min=12206, max=50446, avg=27733.24, stdev=1552.67 00:28:32.867 lat (usec): min=12220, max=50462, avg=27754.95, stdev=1552.07 00:28:32.867 clat percentiles (usec): 00:28:32.867 | 1.00th=[26608], 5.00th=[27395], 10.00th=[27395], 20.00th=[27395], 00:28:32.867 | 30.00th=[27657], 40.00th=[27657], 50.00th=[27657], 60.00th=[27657], 00:28:32.867 | 70.00th=[27657], 80.00th=[27919], 90.00th=[28181], 95.00th=[28443], 00:28:32.867 | 99.00th=[29230], 99.50th=[30016], 99.90th=[50594], 99.95th=[50594], 00:28:32.867 | 99.99th=[50594] 00:28:32.867 bw ( KiB/s): min= 2048, max= 2304, per=4.13%, avg=2282.21, stdev=63.70, samples=19 00:28:32.867 iops : min= 512, max= 576, avg=570.32, stdev=15.86, samples=19 00:28:32.867 lat (msec) : 20=0.28%, 50=99.44%, 100=0.28% 00:28:32.867 cpu : usr=98.73%, sys=0.88%, ctx=15, majf=0, minf=9 00:28:32.867 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.2%, 32=0.0%, >=64=0.0% 00:28:32.867 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.867 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.868 issued rwts: total=5728,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.868 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.868 filename0: (groupid=0, jobs=1): err= 0: pid=766374: Mon Jul 15 18:14:25 2024 00:28:32.868 read: IOPS=572, BW=2288KiB/s (2343kB/s)(22.4MiB/10013msec) 00:28:32.868 slat (nsec): min=7819, max=89297, avg=26734.27, stdev=13416.02 00:28:32.868 clat (usec): min=17894, max=52349, avg=27715.01, stdev=904.32 00:28:32.868 lat (usec): min=17918, max=52374, avg=27741.74, stdev=903.76 00:28:32.868 clat percentiles (usec): 00:28:32.868 | 1.00th=[26870], 5.00th=[27132], 10.00th=[27395], 20.00th=[27395], 00:28:32.868 | 30.00th=[27395], 40.00th=[27657], 50.00th=[27657], 60.00th=[27657], 00:28:32.868 | 70.00th=[27657], 80.00th=[27919], 90.00th=[28181], 95.00th=[28443], 00:28:32.868 | 99.00th=[28967], 99.50th=[29492], 99.90th=[39584], 99.95th=[39584], 00:28:32.868 | 99.99th=[52167] 00:28:32.868 bw ( KiB/s): min= 2176, max= 2304, per=4.13%, avg=2282.21, stdev=47.31, samples=19 00:28:32.868 iops : min= 544, max= 576, avg=570.32, stdev=11.74, samples=19 00:28:32.868 lat (msec) : 20=0.03%, 50=99.93%, 100=0.03% 00:28:32.868 cpu : usr=98.80%, sys=0.81%, ctx=7, majf=0, minf=9 00:28:32.868 IO depths : 1=6.2%, 2=12.4%, 4=25.0%, 8=50.1%, 16=6.3%, 32=0.0%, >=64=0.0% 00:28:32.868 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.868 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.868 issued rwts: total=5728,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.868 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.868 filename0: (groupid=0, jobs=1): err= 0: pid=766375: Mon Jul 15 18:14:25 2024 00:28:32.868 read: IOPS=572, BW=2291KiB/s (2346kB/s)(22.4MiB/10002msec) 00:28:32.868 slat (nsec): min=6058, max=84397, avg=38431.60, stdev=14406.56 00:28:32.868 clat (usec): min=16035, max=38220, avg=27583.76, stdev=939.87 00:28:32.868 lat (usec): min=16042, max=38238, avg=27622.19, stdev=940.93 00:28:32.868 clat percentiles (usec): 00:28:32.868 | 1.00th=[26608], 5.00th=[27132], 10.00th=[27132], 20.00th=[27395], 00:28:32.868 | 30.00th=[27395], 40.00th=[27395], 50.00th=[27657], 60.00th=[27657], 00:28:32.868 | 70.00th=[27657], 80.00th=[27919], 90.00th=[28181], 95.00th=[28443], 00:28:32.868 | 99.00th=[28967], 99.50th=[29230], 99.90th=[38011], 99.95th=[38011], 00:28:32.868 | 99.99th=[38011] 00:28:32.868 bw ( KiB/s): min= 2176, max= 2304, per=4.13%, avg=2282.21, stdev=47.31, samples=19 00:28:32.868 iops : min= 544, max= 576, avg=570.32, stdev=11.74, samples=19 00:28:32.868 lat (msec) : 20=0.28%, 50=99.72% 00:28:32.868 cpu : usr=98.13%, sys=1.09%, ctx=109, majf=0, minf=9 00:28:32.868 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.2%, 32=0.0%, >=64=0.0% 00:28:32.868 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.868 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.868 issued rwts: total=5728,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.868 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.868 filename1: (groupid=0, jobs=1): err= 0: pid=766376: Mon Jul 15 18:14:25 2024 00:28:32.868 read: IOPS=600, BW=2402KiB/s (2460kB/s)(23.5MiB/10015msec) 00:28:32.868 slat (nsec): min=6812, max=78251, avg=20285.29, stdev=11323.53 00:28:32.868 clat (usec): min=10271, max=52900, avg=26468.06, stdev=3490.56 00:28:32.868 lat (usec): min=10280, max=52932, avg=26488.34, stdev=3494.17 00:28:32.868 clat percentiles (usec): 00:28:32.868 | 1.00th=[16057], 5.00th=[18220], 10.00th=[20055], 20.00th=[27132], 00:28:32.868 | 30.00th=[27395], 40.00th=[27395], 50.00th=[27657], 60.00th=[27657], 00:28:32.868 | 70.00th=[27657], 80.00th=[27919], 90.00th=[27919], 95.00th=[28443], 00:28:32.868 | 99.00th=[33817], 99.50th=[37487], 99.90th=[43254], 99.95th=[43254], 00:28:32.868 | 99.99th=[52691] 00:28:32.868 bw ( KiB/s): min= 2160, max= 2912, per=4.35%, avg=2402.63, stdev=210.97, samples=19 00:28:32.868 iops : min= 540, max= 728, avg=600.42, stdev=52.87, samples=19 00:28:32.868 lat (msec) : 20=9.53%, 50=90.44%, 100=0.03% 00:28:32.868 cpu : usr=98.52%, sys=1.09%, ctx=21, majf=0, minf=9 00:28:32.868 IO depths : 1=4.2%, 2=8.8%, 4=19.3%, 8=58.7%, 16=9.0%, 32=0.0%, >=64=0.0% 00:28:32.868 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.868 complete : 0=0.0%, 4=92.7%, 8=2.2%, 16=5.1%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.868 issued rwts: total=6014,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.868 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.868 filename1: (groupid=0, jobs=1): err= 0: pid=766377: Mon Jul 15 18:14:25 2024 00:28:32.868 read: IOPS=572, BW=2291KiB/s (2346kB/s)(22.4MiB/10001msec) 00:28:32.868 slat (nsec): min=4088, max=74617, avg=21241.44, stdev=9084.61 00:28:32.868 clat (usec): min=12333, max=49097, avg=27736.64, stdev=1495.08 00:28:32.868 lat (usec): min=12349, max=49109, avg=27757.88, stdev=1494.40 00:28:32.868 clat percentiles (usec): 00:28:32.868 | 1.00th=[26608], 5.00th=[27395], 10.00th=[27395], 20.00th=[27395], 00:28:32.868 | 30.00th=[27657], 40.00th=[27657], 50.00th=[27657], 60.00th=[27657], 00:28:32.868 | 70.00th=[27657], 80.00th=[27919], 90.00th=[28181], 95.00th=[28443], 00:28:32.868 | 99.00th=[29230], 99.50th=[30016], 99.90th=[49021], 99.95th=[49021], 00:28:32.868 | 99.99th=[49021] 00:28:32.868 bw ( KiB/s): min= 2176, max= 2304, per=4.13%, avg=2282.21, stdev=47.31, samples=19 00:28:32.868 iops : min= 544, max= 576, avg=570.32, stdev=11.74, samples=19 00:28:32.868 lat (msec) : 20=0.28%, 50=99.72% 00:28:32.868 cpu : usr=98.59%, sys=1.02%, ctx=17, majf=0, minf=9 00:28:32.868 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.2%, 32=0.0%, >=64=0.0% 00:28:32.868 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.868 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.868 issued rwts: total=5728,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.868 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.868 filename1: (groupid=0, jobs=1): err= 0: pid=766378: Mon Jul 15 18:14:25 2024 00:28:32.868 read: IOPS=572, BW=2290KiB/s (2345kB/s)(22.4MiB/10006msec) 00:28:32.868 slat (nsec): min=6212, max=93101, avg=26256.34, stdev=13431.32 00:28:32.868 clat (usec): min=12936, max=55379, avg=27700.76, stdev=1464.50 00:28:32.868 lat (usec): min=12949, max=55397, avg=27727.02, stdev=1464.36 00:28:32.868 clat percentiles (usec): 00:28:32.868 | 1.00th=[26608], 5.00th=[27132], 10.00th=[27395], 20.00th=[27395], 00:28:32.868 | 30.00th=[27395], 40.00th=[27657], 50.00th=[27657], 60.00th=[27657], 00:28:32.868 | 70.00th=[27657], 80.00th=[27919], 90.00th=[28181], 95.00th=[28443], 00:28:32.868 | 99.00th=[28967], 99.50th=[29754], 99.90th=[45876], 99.95th=[45876], 00:28:32.868 | 99.99th=[55313] 00:28:32.868 bw ( KiB/s): min= 2176, max= 2304, per=4.13%, avg=2282.42, stdev=46.81, samples=19 00:28:32.868 iops : min= 544, max= 576, avg=570.37, stdev=11.62, samples=19 00:28:32.868 lat (msec) : 20=0.44%, 50=99.53%, 100=0.03% 00:28:32.868 cpu : usr=98.86%, sys=0.77%, ctx=7, majf=0, minf=9 00:28:32.868 IO depths : 1=6.2%, 2=12.4%, 4=24.9%, 8=50.1%, 16=6.3%, 32=0.0%, >=64=0.0% 00:28:32.868 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.868 complete : 0=0.0%, 4=94.1%, 8=0.1%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.868 issued rwts: total=5728,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.868 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.868 filename1: (groupid=0, jobs=1): err= 0: pid=766379: Mon Jul 15 18:14:25 2024 00:28:32.868 read: IOPS=572, BW=2290KiB/s (2345kB/s)(22.4MiB/10004msec) 00:28:32.868 slat (nsec): min=7052, max=96888, avg=45117.22, stdev=20423.63 00:28:32.868 clat (usec): min=19064, max=37071, avg=27498.93, stdev=810.76 00:28:32.868 lat (usec): min=19080, max=37104, avg=27544.05, stdev=813.80 00:28:32.868 clat percentiles (usec): 00:28:32.868 | 1.00th=[26608], 5.00th=[27132], 10.00th=[27132], 20.00th=[27132], 00:28:32.868 | 30.00th=[27132], 40.00th=[27395], 50.00th=[27395], 60.00th=[27657], 00:28:32.868 | 70.00th=[27657], 80.00th=[27657], 90.00th=[27919], 95.00th=[28443], 00:28:32.868 | 99.00th=[28967], 99.50th=[29230], 99.90th=[36963], 99.95th=[36963], 00:28:32.868 | 99.99th=[36963] 00:28:32.868 bw ( KiB/s): min= 2176, max= 2304, per=4.14%, avg=2288.95, stdev=39.87, samples=19 00:28:32.868 iops : min= 544, max= 576, avg=572.00, stdev= 9.91, samples=19 00:28:32.868 lat (msec) : 20=0.28%, 50=99.72% 00:28:32.868 cpu : usr=99.03%, sys=0.59%, ctx=11, majf=0, minf=9 00:28:32.868 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.2%, 32=0.0%, >=64=0.0% 00:28:32.868 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.868 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.868 issued rwts: total=5728,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.868 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.868 filename1: (groupid=0, jobs=1): err= 0: pid=766380: Mon Jul 15 18:14:25 2024 00:28:32.868 read: IOPS=569, BW=2276KiB/s (2331kB/s)(22.2MiB/10005msec) 00:28:32.868 slat (nsec): min=6849, max=93647, avg=28964.19, stdev=20274.66 00:28:32.868 clat (usec): min=6184, max=46709, avg=27978.88, stdev=2731.23 00:28:32.868 lat (usec): min=6199, max=46725, avg=28007.85, stdev=2730.29 00:28:32.868 clat percentiles (usec): 00:28:32.868 | 1.00th=[20579], 5.00th=[27132], 10.00th=[27395], 20.00th=[27657], 00:28:32.868 | 30.00th=[27657], 40.00th=[27657], 50.00th=[27657], 60.00th=[27919], 00:28:32.868 | 70.00th=[27919], 80.00th=[28181], 90.00th=[28443], 95.00th=[29754], 00:28:32.868 | 99.00th=[39060], 99.50th=[42206], 99.90th=[46400], 99.95th=[46924], 00:28:32.868 | 99.99th=[46924] 00:28:32.868 bw ( KiB/s): min= 2144, max= 2308, per=4.10%, avg=2262.47, stdev=51.51, samples=19 00:28:32.868 iops : min= 536, max= 577, avg=565.42, stdev=12.73, samples=19 00:28:32.868 lat (msec) : 10=0.25%, 20=0.69%, 50=99.07% 00:28:32.868 cpu : usr=98.83%, sys=0.79%, ctx=13, majf=0, minf=9 00:28:32.868 IO depths : 1=0.4%, 2=1.2%, 4=5.7%, 8=76.2%, 16=16.6%, 32=0.0%, >=64=0.0% 00:28:32.868 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.868 complete : 0=0.0%, 4=90.5%, 8=7.9%, 16=1.6%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.868 issued rwts: total=5693,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.868 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.868 filename1: (groupid=0, jobs=1): err= 0: pid=766381: Mon Jul 15 18:14:25 2024 00:28:32.868 read: IOPS=571, BW=2288KiB/s (2343kB/s)(22.4MiB/10014msec) 00:28:32.868 slat (nsec): min=7048, max=88383, avg=21188.31, stdev=10280.51 00:28:32.868 clat (usec): min=17162, max=39661, avg=27800.04, stdev=814.06 00:28:32.868 lat (usec): min=17170, max=39686, avg=27821.23, stdev=813.28 00:28:32.868 clat percentiles (usec): 00:28:32.868 | 1.00th=[26870], 5.00th=[27395], 10.00th=[27395], 20.00th=[27657], 00:28:32.868 | 30.00th=[27657], 40.00th=[27657], 50.00th=[27657], 60.00th=[27657], 00:28:32.868 | 70.00th=[27919], 80.00th=[27919], 90.00th=[28443], 95.00th=[28705], 00:28:32.868 | 99.00th=[28967], 99.50th=[29754], 99.90th=[39584], 99.95th=[39584], 00:28:32.868 | 99.99th=[39584] 00:28:32.868 bw ( KiB/s): min= 2176, max= 2304, per=4.13%, avg=2282.21, stdev=47.31, samples=19 00:28:32.868 iops : min= 544, max= 576, avg=570.32, stdev=11.74, samples=19 00:28:32.868 lat (msec) : 20=0.03%, 50=99.97% 00:28:32.868 cpu : usr=98.75%, sys=0.85%, ctx=15, majf=0, minf=9 00:28:32.868 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.3%, 32=0.0%, >=64=0.0% 00:28:32.868 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.868 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.868 issued rwts: total=5728,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.868 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.868 filename1: (groupid=0, jobs=1): err= 0: pid=766382: Mon Jul 15 18:14:25 2024 00:28:32.868 read: IOPS=579, BW=2316KiB/s (2372kB/s)(22.6MiB/10013msec) 00:28:32.868 slat (nsec): min=6836, max=81033, avg=18808.98, stdev=9134.77 00:28:32.868 clat (usec): min=15165, max=39102, avg=27481.25, stdev=1834.96 00:28:32.868 lat (usec): min=15173, max=39134, avg=27500.06, stdev=1836.52 00:28:32.868 clat percentiles (usec): 00:28:32.868 | 1.00th=[17433], 5.00th=[26870], 10.00th=[27395], 20.00th=[27657], 00:28:32.868 | 30.00th=[27657], 40.00th=[27657], 50.00th=[27657], 60.00th=[27657], 00:28:32.868 | 70.00th=[27919], 80.00th=[27919], 90.00th=[28181], 95.00th=[28705], 00:28:32.868 | 99.00th=[29754], 99.50th=[30802], 99.90th=[37487], 99.95th=[38536], 00:28:32.868 | 99.99th=[39060] 00:28:32.868 bw ( KiB/s): min= 2176, max= 2736, per=4.19%, avg=2311.68, stdev=110.16, samples=19 00:28:32.868 iops : min= 544, max= 684, avg=577.68, stdev=27.57, samples=19 00:28:32.868 lat (msec) : 20=2.97%, 50=97.03% 00:28:32.868 cpu : usr=98.90%, sys=0.71%, ctx=10, majf=0, minf=9 00:28:32.868 IO depths : 1=5.8%, 2=11.7%, 4=23.9%, 8=51.8%, 16=6.8%, 32=0.0%, >=64=0.0% 00:28:32.868 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.868 complete : 0=0.0%, 4=93.8%, 8=0.4%, 16=5.8%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.868 issued rwts: total=5798,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.868 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.868 filename1: (groupid=0, jobs=1): err= 0: pid=766383: Mon Jul 15 18:14:25 2024 00:28:32.868 read: IOPS=574, BW=2296KiB/s (2352kB/s)(22.4MiB/10005msec) 00:28:32.868 slat (nsec): min=5346, max=97164, avg=25328.60, stdev=16383.20 00:28:32.868 clat (usec): min=6633, max=60465, avg=27630.95, stdev=2907.13 00:28:32.868 lat (usec): min=6642, max=60484, avg=27656.28, stdev=2907.41 00:28:32.868 clat percentiles (usec): 00:28:32.868 | 1.00th=[16909], 5.00th=[25560], 10.00th=[27132], 20.00th=[27395], 00:28:32.868 | 30.00th=[27395], 40.00th=[27657], 50.00th=[27657], 60.00th=[27657], 00:28:32.868 | 70.00th=[27657], 80.00th=[27919], 90.00th=[28443], 95.00th=[28967], 00:28:32.868 | 99.00th=[40633], 99.50th=[43779], 99.90th=[46400], 99.95th=[46400], 00:28:32.868 | 99.99th=[60556] 00:28:32.868 bw ( KiB/s): min= 2144, max= 2352, per=4.13%, avg=2282.68, stdev=47.23, samples=19 00:28:32.868 iops : min= 536, max= 588, avg=570.47, stdev=11.74, samples=19 00:28:32.868 lat (msec) : 10=0.28%, 20=1.57%, 50=98.12%, 100=0.03% 00:28:32.869 cpu : usr=98.73%, sys=0.89%, ctx=12, majf=0, minf=11 00:28:32.869 IO depths : 1=4.7%, 2=9.9%, 4=21.3%, 8=55.7%, 16=8.5%, 32=0.0%, >=64=0.0% 00:28:32.869 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.869 complete : 0=0.0%, 4=93.3%, 8=1.6%, 16=5.2%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.869 issued rwts: total=5744,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.869 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.869 filename2: (groupid=0, jobs=1): err= 0: pid=766384: Mon Jul 15 18:14:25 2024 00:28:32.869 read: IOPS=572, BW=2288KiB/s (2343kB/s)(22.4MiB/10013msec) 00:28:32.869 slat (nsec): min=8922, max=91609, avg=26163.22, stdev=13276.03 00:28:32.869 clat (usec): min=23711, max=39848, avg=27736.27, stdev=787.26 00:28:32.869 lat (usec): min=23768, max=39882, avg=27762.43, stdev=786.25 00:28:32.869 clat percentiles (usec): 00:28:32.869 | 1.00th=[26870], 5.00th=[27132], 10.00th=[27395], 20.00th=[27395], 00:28:32.869 | 30.00th=[27395], 40.00th=[27657], 50.00th=[27657], 60.00th=[27657], 00:28:32.869 | 70.00th=[27919], 80.00th=[27919], 90.00th=[28181], 95.00th=[28443], 00:28:32.869 | 99.00th=[28967], 99.50th=[29492], 99.90th=[39584], 99.95th=[39584], 00:28:32.869 | 99.99th=[40109] 00:28:32.869 bw ( KiB/s): min= 2176, max= 2304, per=4.13%, avg=2282.21, stdev=47.31, samples=19 00:28:32.869 iops : min= 544, max= 576, avg=570.32, stdev=11.74, samples=19 00:28:32.869 lat (msec) : 50=100.00% 00:28:32.869 cpu : usr=98.85%, sys=0.74%, ctx=12, majf=0, minf=9 00:28:32.869 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.2%, 32=0.0%, >=64=0.0% 00:28:32.869 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.869 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.869 issued rwts: total=5728,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.869 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.869 filename2: (groupid=0, jobs=1): err= 0: pid=766385: Mon Jul 15 18:14:25 2024 00:28:32.869 read: IOPS=574, BW=2298KiB/s (2353kB/s)(22.5MiB/10007msec) 00:28:32.869 slat (nsec): min=6854, max=77968, avg=22138.37, stdev=10777.62 00:28:32.869 clat (usec): min=15110, max=45556, avg=27669.44, stdev=2052.26 00:28:32.869 lat (usec): min=15125, max=45588, avg=27691.58, stdev=2052.47 00:28:32.869 clat percentiles (usec): 00:28:32.869 | 1.00th=[17957], 5.00th=[27132], 10.00th=[27395], 20.00th=[27395], 00:28:32.869 | 30.00th=[27657], 40.00th=[27657], 50.00th=[27657], 60.00th=[27657], 00:28:32.869 | 70.00th=[27919], 80.00th=[27919], 90.00th=[28443], 95.00th=[28705], 00:28:32.869 | 99.00th=[37487], 99.50th=[41157], 99.90th=[45351], 99.95th=[45351], 00:28:32.869 | 99.99th=[45351] 00:28:32.869 bw ( KiB/s): min= 2176, max= 2352, per=4.15%, avg=2291.47, stdev=42.32, samples=19 00:28:32.869 iops : min= 544, max= 588, avg=572.63, stdev=10.54, samples=19 00:28:32.869 lat (msec) : 20=1.67%, 50=98.33% 00:28:32.869 cpu : usr=98.86%, sys=0.75%, ctx=15, majf=0, minf=9 00:28:32.869 IO depths : 1=2.4%, 2=8.3%, 4=24.2%, 8=55.0%, 16=10.1%, 32=0.0%, >=64=0.0% 00:28:32.869 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.869 complete : 0=0.0%, 4=94.1%, 8=0.2%, 16=5.7%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.869 issued rwts: total=5748,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.869 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.869 filename2: (groupid=0, jobs=1): err= 0: pid=766386: Mon Jul 15 18:14:25 2024 00:28:32.869 read: IOPS=575, BW=2303KiB/s (2358kB/s)(22.5MiB/10004msec) 00:28:32.869 slat (nsec): min=3858, max=95134, avg=45091.72, stdev=19601.64 00:28:32.869 clat (usec): min=8462, max=34076, avg=27418.29, stdev=1617.12 00:28:32.869 lat (usec): min=8466, max=34137, avg=27463.38, stdev=1619.17 00:28:32.869 clat percentiles (usec): 00:28:32.869 | 1.00th=[19268], 5.00th=[26870], 10.00th=[27132], 20.00th=[27132], 00:28:32.869 | 30.00th=[27395], 40.00th=[27395], 50.00th=[27657], 60.00th=[27657], 00:28:32.869 | 70.00th=[27657], 80.00th=[27919], 90.00th=[28181], 95.00th=[28443], 00:28:32.869 | 99.00th=[28967], 99.50th=[29230], 99.90th=[29492], 99.95th=[29754], 00:28:32.869 | 99.99th=[33817] 00:28:32.869 bw ( KiB/s): min= 2176, max= 2436, per=4.17%, avg=2302.63, stdev=43.41, samples=19 00:28:32.869 iops : min= 544, max= 609, avg=575.42, stdev=10.88, samples=19 00:28:32.869 lat (msec) : 10=0.56%, 20=0.56%, 50=98.89% 00:28:32.869 cpu : usr=98.71%, sys=0.90%, ctx=15, majf=0, minf=9 00:28:32.869 IO depths : 1=6.2%, 2=12.4%, 4=25.0%, 8=50.1%, 16=6.3%, 32=0.0%, >=64=0.0% 00:28:32.869 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.869 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.869 issued rwts: total=5760,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.869 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.869 filename2: (groupid=0, jobs=1): err= 0: pid=766387: Mon Jul 15 18:14:25 2024 00:28:32.869 read: IOPS=580, BW=2323KiB/s (2379kB/s)(22.7MiB/10005msec) 00:28:32.869 slat (usec): min=4, max=104, avg=24.93, stdev=15.69 00:28:32.869 clat (usec): min=6043, max=60043, avg=27311.78, stdev=2669.10 00:28:32.869 lat (usec): min=6057, max=60060, avg=27336.71, stdev=2670.76 00:28:32.869 clat percentiles (usec): 00:28:32.869 | 1.00th=[18744], 5.00th=[22676], 10.00th=[27132], 20.00th=[27395], 00:28:32.869 | 30.00th=[27395], 40.00th=[27657], 50.00th=[27657], 60.00th=[27657], 00:28:32.869 | 70.00th=[27657], 80.00th=[27919], 90.00th=[28181], 95.00th=[28443], 00:28:32.869 | 99.00th=[33162], 99.50th=[38011], 99.90th=[46400], 99.95th=[60031], 00:28:32.869 | 99.99th=[60031] 00:28:32.869 bw ( KiB/s): min= 2048, max= 2592, per=4.18%, avg=2310.47, stdev=93.48, samples=19 00:28:32.869 iops : min= 512, max= 648, avg=577.42, stdev=23.40, samples=19 00:28:32.869 lat (msec) : 10=0.28%, 20=2.70%, 50=96.97%, 100=0.05% 00:28:32.869 cpu : usr=98.81%, sys=0.81%, ctx=15, majf=0, minf=9 00:28:32.869 IO depths : 1=5.5%, 2=11.1%, 4=22.8%, 8=53.4%, 16=7.2%, 32=0.0%, >=64=0.0% 00:28:32.869 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.869 complete : 0=0.0%, 4=93.5%, 8=0.9%, 16=5.6%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.869 issued rwts: total=5810,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.869 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.869 filename2: (groupid=0, jobs=1): err= 0: pid=766388: Mon Jul 15 18:14:25 2024 00:28:32.869 read: IOPS=572, BW=2290KiB/s (2345kB/s)(22.4MiB/10005msec) 00:28:32.869 slat (nsec): min=6146, max=89138, avg=25974.05, stdev=13404.12 00:28:32.869 clat (usec): min=13059, max=43598, avg=27689.54, stdev=1219.07 00:28:32.869 lat (usec): min=13073, max=43615, avg=27715.51, stdev=1218.57 00:28:32.869 clat percentiles (usec): 00:28:32.869 | 1.00th=[26870], 5.00th=[27132], 10.00th=[27395], 20.00th=[27395], 00:28:32.869 | 30.00th=[27395], 40.00th=[27657], 50.00th=[27657], 60.00th=[27657], 00:28:32.869 | 70.00th=[27657], 80.00th=[27919], 90.00th=[28181], 95.00th=[28443], 00:28:32.869 | 99.00th=[28967], 99.50th=[29492], 99.90th=[43779], 99.95th=[43779], 00:28:32.869 | 99.99th=[43779] 00:28:32.869 bw ( KiB/s): min= 2176, max= 2304, per=4.13%, avg=2282.21, stdev=47.31, samples=19 00:28:32.869 iops : min= 544, max= 576, avg=570.32, stdev=11.74, samples=19 00:28:32.869 lat (msec) : 20=0.28%, 50=99.72% 00:28:32.869 cpu : usr=98.56%, sys=1.06%, ctx=15, majf=0, minf=9 00:28:32.869 IO depths : 1=6.2%, 2=12.5%, 4=25.0%, 8=50.0%, 16=6.2%, 32=0.0%, >=64=0.0% 00:28:32.869 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.869 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.869 issued rwts: total=5728,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.869 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.869 filename2: (groupid=0, jobs=1): err= 0: pid=766389: Mon Jul 15 18:14:25 2024 00:28:32.869 read: IOPS=575, BW=2301KiB/s (2357kB/s)(22.5MiB/10011msec) 00:28:32.869 slat (nsec): min=3239, max=94452, avg=30099.47, stdev=21001.27 00:28:32.869 clat (usec): min=8063, max=29593, avg=27581.24, stdev=1431.42 00:28:32.869 lat (usec): min=8073, max=29610, avg=27611.34, stdev=1429.91 00:28:32.869 clat percentiles (usec): 00:28:32.869 | 1.00th=[19268], 5.00th=[27132], 10.00th=[27132], 20.00th=[27395], 00:28:32.869 | 30.00th=[27395], 40.00th=[27657], 50.00th=[27657], 60.00th=[27657], 00:28:32.869 | 70.00th=[27919], 80.00th=[27919], 90.00th=[28443], 95.00th=[28443], 00:28:32.869 | 99.00th=[28967], 99.50th=[29230], 99.90th=[29492], 99.95th=[29492], 00:28:32.869 | 99.99th=[29492] 00:28:32.869 bw ( KiB/s): min= 2171, max= 2432, per=4.16%, avg=2295.85, stdev=65.20, samples=20 00:28:32.869 iops : min= 542, max= 608, avg=573.70, stdev=16.29, samples=20 00:28:32.869 lat (msec) : 10=0.24%, 20=0.87%, 50=98.89% 00:28:32.869 cpu : usr=99.07%, sys=0.54%, ctx=9, majf=0, minf=9 00:28:32.869 IO depths : 1=6.2%, 2=12.4%, 4=24.9%, 8=50.2%, 16=6.3%, 32=0.0%, >=64=0.0% 00:28:32.869 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.869 complete : 0=0.0%, 4=94.1%, 8=0.1%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.869 issued rwts: total=5760,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.869 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.869 filename2: (groupid=0, jobs=1): err= 0: pid=766390: Mon Jul 15 18:14:25 2024 00:28:32.869 read: IOPS=575, BW=2302KiB/s (2357kB/s)(22.5MiB/10010msec) 00:28:32.869 slat (nsec): min=4233, max=97095, avg=38012.99, stdev=22068.23 00:28:32.869 clat (usec): min=9449, max=31240, avg=27510.97, stdev=1464.10 00:28:32.869 lat (usec): min=9463, max=31298, avg=27548.98, stdev=1464.02 00:28:32.869 clat percentiles (usec): 00:28:32.869 | 1.00th=[19268], 5.00th=[26870], 10.00th=[27132], 20.00th=[27395], 00:28:32.869 | 30.00th=[27395], 40.00th=[27657], 50.00th=[27657], 60.00th=[27657], 00:28:32.869 | 70.00th=[27919], 80.00th=[27919], 90.00th=[28181], 95.00th=[28443], 00:28:32.869 | 99.00th=[28967], 99.50th=[29230], 99.90th=[29230], 99.95th=[29492], 00:28:32.869 | 99.99th=[31327] 00:28:32.869 bw ( KiB/s): min= 2171, max= 2436, per=4.16%, avg=2296.05, stdev=65.64, samples=20 00:28:32.869 iops : min= 542, max= 609, avg=573.75, stdev=16.40, samples=20 00:28:32.869 lat (msec) : 10=0.24%, 20=0.87%, 50=98.89% 00:28:32.869 cpu : usr=98.98%, sys=0.62%, ctx=11, majf=0, minf=9 00:28:32.869 IO depths : 1=6.2%, 2=12.4%, 4=25.0%, 8=50.1%, 16=6.3%, 32=0.0%, >=64=0.0% 00:28:32.869 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.869 complete : 0=0.0%, 4=94.1%, 8=0.0%, 16=5.9%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.869 issued rwts: total=5760,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.869 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.869 filename2: (groupid=0, jobs=1): err= 0: pid=766391: Mon Jul 15 18:14:25 2024 00:28:32.869 read: IOPS=568, BW=2276KiB/s (2330kB/s)(22.2MiB/10005msec) 00:28:32.869 slat (usec): min=6, max=101, avg=32.13, stdev=16.47 00:28:32.869 clat (usec): min=6407, max=55867, avg=27931.74, stdev=3234.20 00:28:32.869 lat (usec): min=6426, max=55899, avg=27963.87, stdev=3233.80 00:28:32.869 clat percentiles (usec): 00:28:32.869 | 1.00th=[19530], 5.00th=[26870], 10.00th=[27395], 20.00th=[27657], 00:28:32.869 | 30.00th=[27657], 40.00th=[27657], 50.00th=[27657], 60.00th=[27657], 00:28:32.869 | 70.00th=[27919], 80.00th=[27919], 90.00th=[28443], 95.00th=[34341], 00:28:32.869 | 99.00th=[38536], 99.50th=[40633], 99.90th=[55837], 99.95th=[55837], 00:28:32.869 | 99.99th=[55837] 00:28:32.869 bw ( KiB/s): min= 2052, max= 2320, per=4.09%, avg=2261.00, stdev=69.88, samples=19 00:28:32.869 iops : min= 513, max= 580, avg=565.05, stdev=17.36, samples=19 00:28:32.869 lat (msec) : 10=0.33%, 20=2.30%, 50=97.08%, 100=0.28% 00:28:32.869 cpu : usr=96.81%, sys=1.89%, ctx=218, majf=0, minf=9 00:28:32.869 IO depths : 1=0.5%, 2=2.7%, 4=11.0%, 8=70.5%, 16=15.3%, 32=0.0%, >=64=0.0% 00:28:32.869 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.869 complete : 0=0.0%, 4=91.6%, 8=5.8%, 16=2.7%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:32.869 issued rwts: total=5692,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:32.869 latency : target=0, window=0, percentile=100.00%, depth=16 00:28:32.869 00:28:32.869 Run status group 0 (all jobs): 00:28:32.869 READ: bw=53.9MiB/s (56.5MB/s), 2276KiB/s-2402KiB/s (2330kB/s-2460kB/s), io=540MiB (566MB), run=10001-10015msec 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@113 -- # destroy_subsystems 0 1 2 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@43 -- # local sub 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 0 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=0 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 1 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=1 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null1 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:28:32.869 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 2 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=2 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode2 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null2 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # NULL_DIF=1 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # bs=8k,16k,128k 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # numjobs=2 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # iodepth=8 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # runtime=5 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@115 -- # files=1 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@117 -- # create_subsystems 0 1 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@28 -- # local sub 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 0 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=0 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 1 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:32.870 bdev_null0 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:32.870 [2024-07-15 18:14:25.428105] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@30 -- # for sub in "$@" 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@31 -- # create_subsystem 1 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@18 -- # local sub_id=1 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null1 64 512 --md-size 16 --dif-type 1 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:32.870 bdev_null1 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode1 --serial-number 53313233-1 --allow-any-host 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode1 bdev_null1 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode1 -t tcp -a 10.0.0.2 -s 4420 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@118 -- # fio /dev/fd/62 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@118 -- # create_json_sub_conf 0 1 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@51 -- # gen_nvmf_target_json 0 1 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@532 -- # config=() 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@532 -- # local subsystem config 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@82 -- # gen_fio_conf 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@54 -- # local file 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:28:32.870 { 00:28:32.870 "params": { 00:28:32.870 "name": "Nvme$subsystem", 00:28:32.870 "trtype": "$TEST_TRANSPORT", 00:28:32.870 "traddr": "$NVMF_FIRST_TARGET_IP", 00:28:32.870 "adrfam": "ipv4", 00:28:32.870 "trsvcid": "$NVMF_PORT", 00:28:32.870 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:28:32.870 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:28:32.870 "hdgst": ${hdgst:-false}, 00:28:32.870 "ddgst": ${ddgst:-false} 00:28:32.870 }, 00:28:32.870 "method": "bdev_nvme_attach_controller" 00:28:32.870 } 00:28:32.870 EOF 00:28:32.870 )") 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@56 -- # cat 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1339 -- # local sanitizers 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1341 -- # shift 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1343 -- # local asan_lib= 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@554 -- # cat 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file = 1 )) 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@73 -- # cat 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # grep libasan 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:28:32.870 { 00:28:32.870 "params": { 00:28:32.870 "name": "Nvme$subsystem", 00:28:32.870 "trtype": "$TEST_TRANSPORT", 00:28:32.870 "traddr": "$NVMF_FIRST_TARGET_IP", 00:28:32.870 "adrfam": "ipv4", 00:28:32.870 "trsvcid": "$NVMF_PORT", 00:28:32.870 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:28:32.870 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:28:32.870 "hdgst": ${hdgst:-false}, 00:28:32.870 "ddgst": ${ddgst:-false} 00:28:32.870 }, 00:28:32.870 "method": "bdev_nvme_attach_controller" 00:28:32.870 } 00:28:32.870 EOF 00:28:32.870 )") 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file++ )) 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- target/dif.sh@72 -- # (( file <= files )) 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@554 -- # cat 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@556 -- # jq . 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@557 -- # IFS=, 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:28:32.870 "params": { 00:28:32.870 "name": "Nvme0", 00:28:32.870 "trtype": "tcp", 00:28:32.870 "traddr": "10.0.0.2", 00:28:32.870 "adrfam": "ipv4", 00:28:32.870 "trsvcid": "4420", 00:28:32.870 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:28:32.870 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:28:32.870 "hdgst": false, 00:28:32.870 "ddgst": false 00:28:32.870 }, 00:28:32.870 "method": "bdev_nvme_attach_controller" 00:28:32.870 },{ 00:28:32.870 "params": { 00:28:32.870 "name": "Nvme1", 00:28:32.870 "trtype": "tcp", 00:28:32.870 "traddr": "10.0.0.2", 00:28:32.870 "adrfam": "ipv4", 00:28:32.870 "trsvcid": "4420", 00:28:32.870 "subnqn": "nqn.2016-06.io.spdk:cnode1", 00:28:32.870 "hostnqn": "nqn.2016-06.io.spdk:host1", 00:28:32.870 "hdgst": false, 00:28:32.870 "ddgst": false 00:28:32.870 }, 00:28:32.870 "method": "bdev_nvme_attach_controller" 00:28:32.870 }' 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # asan_lib= 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1345 -- # asan_lib= 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:28:32.870 18:14:25 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:28:32.870 filename0: (g=0): rw=randread, bs=(R) 8192B-8192B, (W) 16.0KiB-16.0KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=8 00:28:32.870 ... 00:28:32.870 filename1: (g=0): rw=randread, bs=(R) 8192B-8192B, (W) 16.0KiB-16.0KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=8 00:28:32.870 ... 00:28:32.870 fio-3.35 00:28:32.870 Starting 4 threads 00:28:32.870 EAL: No free 2048 kB hugepages reported on node 1 00:28:38.144 00:28:38.144 filename0: (groupid=0, jobs=1): err= 0: pid=768337: Mon Jul 15 18:14:31 2024 00:28:38.144 read: IOPS=2695, BW=21.1MiB/s (22.1MB/s)(105MiB/5003msec) 00:28:38.145 slat (usec): min=6, max=188, avg= 9.07, stdev= 3.27 00:28:38.145 clat (usec): min=1400, max=43414, avg=2941.16, stdev=1097.24 00:28:38.145 lat (usec): min=1407, max=43439, avg=2950.23, stdev=1097.27 00:28:38.145 clat percentiles (usec): 00:28:38.145 | 1.00th=[ 1876], 5.00th=[ 2245], 10.00th=[ 2409], 20.00th=[ 2606], 00:28:38.145 | 30.00th=[ 2737], 40.00th=[ 2802], 50.00th=[ 2900], 60.00th=[ 2966], 00:28:38.145 | 70.00th=[ 2999], 80.00th=[ 3032], 90.00th=[ 3425], 95.00th=[ 4080], 00:28:38.145 | 99.00th=[ 4490], 99.50th=[ 4686], 99.90th=[ 5080], 99.95th=[43254], 00:28:38.145 | 99.99th=[43254] 00:28:38.145 bw ( KiB/s): min=19632, max=22288, per=25.64%, avg=21564.80, stdev=848.14, samples=10 00:28:38.145 iops : min= 2454, max= 2786, avg=2695.60, stdev=106.02, samples=10 00:28:38.145 lat (msec) : 2=1.60%, 4=92.95%, 10=5.39%, 50=0.06% 00:28:38.145 cpu : usr=96.00%, sys=3.68%, ctx=11, majf=0, minf=15 00:28:38.145 IO depths : 1=0.4%, 2=4.0%, 4=68.1%, 8=27.5%, 16=0.0%, 32=0.0%, >=64=0.0% 00:28:38.145 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:38.145 complete : 0=0.0%, 4=92.6%, 8=7.4%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:38.145 issued rwts: total=13484,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:38.145 latency : target=0, window=0, percentile=100.00%, depth=8 00:28:38.145 filename0: (groupid=0, jobs=1): err= 0: pid=768338: Mon Jul 15 18:14:31 2024 00:28:38.145 read: IOPS=2585, BW=20.2MiB/s (21.2MB/s)(101MiB/5001msec) 00:28:38.145 slat (nsec): min=6192, max=29116, avg=9151.20, stdev=2967.20 00:28:38.145 clat (usec): min=895, max=45108, avg=3068.47, stdev=1145.33 00:28:38.145 lat (usec): min=902, max=45129, avg=3077.62, stdev=1145.24 00:28:38.145 clat percentiles (usec): 00:28:38.145 | 1.00th=[ 2147], 5.00th=[ 2507], 10.00th=[ 2638], 20.00th=[ 2769], 00:28:38.145 | 30.00th=[ 2835], 40.00th=[ 2900], 50.00th=[ 2966], 60.00th=[ 2999], 00:28:38.145 | 70.00th=[ 3032], 80.00th=[ 3228], 90.00th=[ 3621], 95.00th=[ 4113], 00:28:38.145 | 99.00th=[ 4686], 99.50th=[ 4817], 99.90th=[ 5342], 99.95th=[44827], 00:28:38.145 | 99.99th=[45351] 00:28:38.145 bw ( KiB/s): min=19334, max=21344, per=24.59%, avg=20686.89, stdev=609.61, samples=9 00:28:38.145 iops : min= 2416, max= 2668, avg=2585.78, stdev=76.41, samples=9 00:28:38.145 lat (usec) : 1000=0.02% 00:28:38.145 lat (msec) : 2=0.45%, 4=93.24%, 10=6.23%, 50=0.06% 00:28:38.145 cpu : usr=96.64%, sys=3.04%, ctx=10, majf=0, minf=42 00:28:38.145 IO depths : 1=0.2%, 2=1.6%, 4=69.7%, 8=28.6%, 16=0.0%, 32=0.0%, >=64=0.0% 00:28:38.145 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:38.145 complete : 0=0.0%, 4=93.4%, 8=6.6%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:38.145 issued rwts: total=12931,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:38.145 latency : target=0, window=0, percentile=100.00%, depth=8 00:28:38.145 filename1: (groupid=0, jobs=1): err= 0: pid=768339: Mon Jul 15 18:14:31 2024 00:28:38.145 read: IOPS=2639, BW=20.6MiB/s (21.6MB/s)(103MiB/5001msec) 00:28:38.145 slat (nsec): min=6214, max=86604, avg=9110.46, stdev=2948.63 00:28:38.145 clat (usec): min=1332, max=42820, avg=3006.02, stdev=1096.82 00:28:38.145 lat (usec): min=1342, max=42838, avg=3015.13, stdev=1096.73 00:28:38.145 clat percentiles (usec): 00:28:38.145 | 1.00th=[ 2024], 5.00th=[ 2311], 10.00th=[ 2507], 20.00th=[ 2671], 00:28:38.145 | 30.00th=[ 2802], 40.00th=[ 2868], 50.00th=[ 2933], 60.00th=[ 2999], 00:28:38.145 | 70.00th=[ 3032], 80.00th=[ 3130], 90.00th=[ 3589], 95.00th=[ 4113], 00:28:38.145 | 99.00th=[ 4752], 99.50th=[ 4817], 99.90th=[ 5014], 99.95th=[42730], 00:28:38.145 | 99.99th=[42730] 00:28:38.145 bw ( KiB/s): min=20064, max=22160, per=25.00%, avg=21031.11, stdev=656.44, samples=9 00:28:38.145 iops : min= 2508, max= 2770, avg=2628.89, stdev=82.06, samples=9 00:28:38.145 lat (msec) : 2=0.86%, 4=92.70%, 10=6.38%, 50=0.06% 00:28:38.145 cpu : usr=96.32%, sys=3.36%, ctx=10, majf=0, minf=69 00:28:38.145 IO depths : 1=0.2%, 2=2.3%, 4=68.9%, 8=28.6%, 16=0.0%, 32=0.0%, >=64=0.0% 00:28:38.145 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:38.145 complete : 0=0.0%, 4=93.4%, 8=6.6%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:38.145 issued rwts: total=13198,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:38.145 latency : target=0, window=0, percentile=100.00%, depth=8 00:28:38.145 filename1: (groupid=0, jobs=1): err= 0: pid=768340: Mon Jul 15 18:14:31 2024 00:28:38.145 read: IOPS=2597, BW=20.3MiB/s (21.3MB/s)(101MiB/5001msec) 00:28:38.145 slat (nsec): min=6206, max=29298, avg=8684.47, stdev=2701.80 00:28:38.145 clat (usec): min=988, max=5564, avg=3055.82, stdev=598.32 00:28:38.145 lat (usec): min=995, max=5571, avg=3064.50, stdev=597.90 00:28:38.145 clat percentiles (usec): 00:28:38.145 | 1.00th=[ 1037], 5.00th=[ 2442], 10.00th=[ 2606], 20.00th=[ 2737], 00:28:38.145 | 30.00th=[ 2835], 40.00th=[ 2900], 50.00th=[ 2966], 60.00th=[ 2999], 00:28:38.145 | 70.00th=[ 3032], 80.00th=[ 3294], 90.00th=[ 4047], 95.00th=[ 4359], 00:28:38.145 | 99.00th=[ 4752], 99.50th=[ 4817], 99.90th=[ 5145], 99.95th=[ 5276], 00:28:38.145 | 99.99th=[ 5538] 00:28:38.145 bw ( KiB/s): min=19952, max=22864, per=24.75%, avg=20816.00, stdev=852.44, samples=9 00:28:38.145 iops : min= 2494, max= 2858, avg=2602.00, stdev=106.56, samples=9 00:28:38.145 lat (usec) : 1000=0.05% 00:28:38.145 lat (msec) : 2=2.21%, 4=87.38%, 10=10.35% 00:28:38.145 cpu : usr=96.48%, sys=3.22%, ctx=6, majf=0, minf=38 00:28:38.145 IO depths : 1=0.1%, 2=1.3%, 4=70.5%, 8=28.2%, 16=0.0%, 32=0.0%, >=64=0.0% 00:28:38.145 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:38.145 complete : 0=0.0%, 4=93.2%, 8=6.8%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:38.145 issued rwts: total=12990,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:38.145 latency : target=0, window=0, percentile=100.00%, depth=8 00:28:38.145 00:28:38.145 Run status group 0 (all jobs): 00:28:38.145 READ: bw=82.1MiB/s (86.1MB/s), 20.2MiB/s-21.1MiB/s (21.2MB/s-22.1MB/s), io=411MiB (431MB), run=5001-5003msec 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- target/dif.sh@119 -- # destroy_subsystems 0 1 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- target/dif.sh@43 -- # local sub 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 0 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=0 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- target/dif.sh@45 -- # for sub in "$@" 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- target/dif.sh@46 -- # destroy_subsystem 1 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- target/dif.sh@36 -- # local sub_id=1 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null1 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:38.145 00:28:38.145 real 0m24.410s 00:28:38.145 user 4m52.301s 00:28:38.145 sys 0m4.353s 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@1124 -- # xtrace_disable 00:28:38.145 18:14:31 nvmf_dif.fio_dif_rand_params -- common/autotest_common.sh@10 -- # set +x 00:28:38.145 ************************************ 00:28:38.145 END TEST fio_dif_rand_params 00:28:38.145 ************************************ 00:28:38.145 18:14:31 nvmf_dif -- common/autotest_common.sh@1142 -- # return 0 00:28:38.145 18:14:31 nvmf_dif -- target/dif.sh@144 -- # run_test fio_dif_digest fio_dif_digest 00:28:38.145 18:14:31 nvmf_dif -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:28:38.145 18:14:31 nvmf_dif -- common/autotest_common.sh@1105 -- # xtrace_disable 00:28:38.145 18:14:31 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:28:38.145 ************************************ 00:28:38.145 START TEST fio_dif_digest 00:28:38.145 ************************************ 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1123 -- # fio_dif_digest 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@123 -- # local NULL_DIF 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@124 -- # local bs numjobs runtime iodepth files 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@125 -- # local hdgst ddgst 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # NULL_DIF=3 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # bs=128k,128k,128k 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # numjobs=3 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # iodepth=3 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@127 -- # runtime=10 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@128 -- # hdgst=true 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@128 -- # ddgst=true 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@130 -- # create_subsystems 0 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@28 -- # local sub 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@30 -- # for sub in "$@" 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@31 -- # create_subsystem 0 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@18 -- # local sub_id=0 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@21 -- # rpc_cmd bdev_null_create bdev_null0 64 512 --md-size 16 --dif-type 3 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:28:38.145 bdev_null0 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@22 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:cnode0 --serial-number 53313233-0 --allow-any-host 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@23 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:cnode0 bdev_null0 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@24 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:cnode0 -t tcp -a 10.0.0.2 -s 4420 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:38.145 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:28:38.146 [2024-07-15 18:14:31.844647] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@131 -- # fio /dev/fd/62 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@131 -- # create_json_sub_conf 0 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@51 -- # gen_nvmf_target_json 0 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@82 -- # gen_fio_conf 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- nvmf/common.sh@532 -- # config=() 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@82 -- # fio_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@54 -- # local file 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@56 -- # cat 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- nvmf/common.sh@532 -- # local subsystem config 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1356 -- # fio_plugin /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- nvmf/common.sh@534 -- # for subsystem in "${@:-1}" 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1337 -- # local fio_dir=/usr/src/fio 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- nvmf/common.sh@554 -- # config+=("$(cat <<-EOF 00:28:38.146 { 00:28:38.146 "params": { 00:28:38.146 "name": "Nvme$subsystem", 00:28:38.146 "trtype": "$TEST_TRANSPORT", 00:28:38.146 "traddr": "$NVMF_FIRST_TARGET_IP", 00:28:38.146 "adrfam": "ipv4", 00:28:38.146 "trsvcid": "$NVMF_PORT", 00:28:38.146 "subnqn": "nqn.2016-06.io.spdk:cnode$subsystem", 00:28:38.146 "hostnqn": "nqn.2016-06.io.spdk:host$subsystem", 00:28:38.146 "hdgst": ${hdgst:-false}, 00:28:38.146 "ddgst": ${ddgst:-false} 00:28:38.146 }, 00:28:38.146 "method": "bdev_nvme_attach_controller" 00:28:38.146 } 00:28:38.146 EOF 00:28:38.146 )") 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1339 -- # sanitizers=('libasan' 'libclang_rt.asan') 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1339 -- # local sanitizers 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1340 -- # local plugin=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- nvmf/common.sh@554 -- # cat 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1341 -- # shift 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1343 -- # local asan_lib= 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@72 -- # (( file = 1 )) 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- target/dif.sh@72 -- # (( file <= files )) 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- nvmf/common.sh@556 -- # jq . 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # grep libasan 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- nvmf/common.sh@557 -- # IFS=, 00:28:38.146 18:14:31 nvmf_dif.fio_dif_digest -- nvmf/common.sh@558 -- # printf '%s\n' '{ 00:28:38.146 "params": { 00:28:38.146 "name": "Nvme0", 00:28:38.146 "trtype": "tcp", 00:28:38.146 "traddr": "10.0.0.2", 00:28:38.146 "adrfam": "ipv4", 00:28:38.146 "trsvcid": "4420", 00:28:38.146 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:28:38.146 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:28:38.146 "hdgst": true, 00:28:38.146 "ddgst": true 00:28:38.146 }, 00:28:38.146 "method": "bdev_nvme_attach_controller" 00:28:38.146 }' 00:28:38.433 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # asan_lib= 00:28:38.433 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:28:38.433 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1344 -- # for sanitizer in "${sanitizers[@]}" 00:28:38.433 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # ldd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev 00:28:38.433 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # grep libclang_rt.asan 00:28:38.433 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # awk '{print $3}' 00:28:38.433 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1345 -- # asan_lib= 00:28:38.433 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1346 -- # [[ -n '' ]] 00:28:38.433 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1352 -- # LD_PRELOAD=' /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/fio/spdk_bdev' 00:28:38.433 18:14:31 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1352 -- # /usr/src/fio/fio --ioengine=spdk_bdev --spdk_json_conf /dev/fd/62 /dev/fd/61 00:28:38.697 filename0: (g=0): rw=randread, bs=(R) 128KiB-128KiB, (W) 128KiB-128KiB, (T) 128KiB-128KiB, ioengine=spdk_bdev, iodepth=3 00:28:38.697 ... 00:28:38.697 fio-3.35 00:28:38.697 Starting 3 threads 00:28:38.697 EAL: No free 2048 kB hugepages reported on node 1 00:28:50.961 00:28:50.961 filename0: (groupid=0, jobs=1): err= 0: pid=769573: Mon Jul 15 18:14:42 2024 00:28:50.961 read: IOPS=275, BW=34.4MiB/s (36.1MB/s)(346MiB/10044msec) 00:28:50.961 slat (nsec): min=4570, max=71656, avg=11192.47, stdev=2580.88 00:28:50.961 clat (usec): min=6382, max=54448, avg=10852.88, stdev=1839.40 00:28:50.961 lat (usec): min=6391, max=54456, avg=10864.07, stdev=1839.39 00:28:50.961 clat percentiles (usec): 00:28:50.961 | 1.00th=[ 8356], 5.00th=[ 9241], 10.00th=[ 9765], 20.00th=[10028], 00:28:50.961 | 30.00th=[10421], 40.00th=[10552], 50.00th=[10814], 60.00th=[10945], 00:28:50.961 | 70.00th=[11207], 80.00th=[11469], 90.00th=[11994], 95.00th=[12256], 00:28:50.961 | 99.00th=[13042], 99.50th=[13435], 99.90th=[52167], 99.95th=[53740], 00:28:50.961 | 99.99th=[54264] 00:28:50.961 bw ( KiB/s): min=31488, max=37376, per=33.89%, avg=35379.20, stdev=1378.85, samples=20 00:28:50.961 iops : min= 246, max= 292, avg=276.40, stdev=10.77, samples=20 00:28:50.961 lat (msec) : 10=17.65%, 20=82.21%, 50=0.04%, 100=0.11% 00:28:50.961 cpu : usr=94.37%, sys=5.31%, ctx=19, majf=0, minf=82 00:28:50.961 IO depths : 1=0.6%, 2=99.4%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:28:50.961 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:50.961 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:50.961 issued rwts: total=2765,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:50.961 latency : target=0, window=0, percentile=100.00%, depth=3 00:28:50.961 filename0: (groupid=0, jobs=1): err= 0: pid=769574: Mon Jul 15 18:14:42 2024 00:28:50.961 read: IOPS=276, BW=34.5MiB/s (36.2MB/s)(347MiB/10045msec) 00:28:50.961 slat (nsec): min=6436, max=34910, avg=11238.94, stdev=2375.96 00:28:50.961 clat (usec): min=6161, max=49033, avg=10830.33, stdev=1341.46 00:28:50.961 lat (usec): min=6169, max=49040, avg=10841.57, stdev=1341.51 00:28:50.961 clat percentiles (usec): 00:28:50.961 | 1.00th=[ 7898], 5.00th=[ 9372], 10.00th=[ 9765], 20.00th=[10159], 00:28:50.961 | 30.00th=[10421], 40.00th=[10552], 50.00th=[10814], 60.00th=[11076], 00:28:50.961 | 70.00th=[11207], 80.00th=[11469], 90.00th=[11994], 95.00th=[12256], 00:28:50.961 | 99.00th=[13042], 99.50th=[13304], 99.90th=[14746], 99.95th=[45351], 00:28:50.961 | 99.99th=[49021] 00:28:50.961 bw ( KiB/s): min=34304, max=36608, per=34.00%, avg=35494.40, stdev=744.16, samples=20 00:28:50.961 iops : min= 268, max= 286, avg=277.30, stdev= 5.81, samples=20 00:28:50.961 lat (msec) : 10=16.04%, 20=83.89%, 50=0.07% 00:28:50.961 cpu : usr=94.28%, sys=5.42%, ctx=21, majf=0, minf=154 00:28:50.961 IO depths : 1=0.4%, 2=99.6%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:28:50.961 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:50.961 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:50.961 issued rwts: total=2775,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:50.961 latency : target=0, window=0, percentile=100.00%, depth=3 00:28:50.961 filename0: (groupid=0, jobs=1): err= 0: pid=769575: Mon Jul 15 18:14:42 2024 00:28:50.961 read: IOPS=264, BW=33.0MiB/s (34.6MB/s)(332MiB/10046msec) 00:28:50.961 slat (nsec): min=6479, max=55603, avg=11190.27, stdev=2555.52 00:28:50.961 clat (usec): min=6657, max=53143, avg=11329.19, stdev=2354.88 00:28:50.961 lat (usec): min=6669, max=53155, avg=11340.39, stdev=2354.88 00:28:50.961 clat percentiles (usec): 00:28:50.961 | 1.00th=[ 8979], 5.00th=[ 9765], 10.00th=[10159], 20.00th=[10552], 00:28:50.961 | 30.00th=[10814], 40.00th=[10945], 50.00th=[11207], 60.00th=[11338], 00:28:50.961 | 70.00th=[11600], 80.00th=[11994], 90.00th=[12387], 95.00th=[12780], 00:28:50.961 | 99.00th=[13698], 99.50th=[14091], 99.90th=[51643], 99.95th=[53216], 00:28:50.961 | 99.99th=[53216] 00:28:50.961 bw ( KiB/s): min=30720, max=37120, per=32.50%, avg=33932.80, stdev=1277.23, samples=20 00:28:50.961 iops : min= 240, max= 290, avg=265.10, stdev= 9.98, samples=20 00:28:50.961 lat (msec) : 10=8.25%, 20=91.44%, 50=0.08%, 100=0.23% 00:28:50.961 cpu : usr=94.65%, sys=5.03%, ctx=21, majf=0, minf=135 00:28:50.961 IO depths : 1=0.9%, 2=99.1%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0% 00:28:50.961 submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:50.961 complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0% 00:28:50.961 issued rwts: total=2653,0,0,0 short=0,0,0,0 dropped=0,0,0,0 00:28:50.961 latency : target=0, window=0, percentile=100.00%, depth=3 00:28:50.961 00:28:50.961 Run status group 0 (all jobs): 00:28:50.961 READ: bw=102MiB/s (107MB/s), 33.0MiB/s-34.5MiB/s (34.6MB/s-36.2MB/s), io=1024MiB (1074MB), run=10044-10046msec 00:28:50.961 18:14:42 nvmf_dif.fio_dif_digest -- target/dif.sh@132 -- # destroy_subsystems 0 00:28:50.961 18:14:42 nvmf_dif.fio_dif_digest -- target/dif.sh@43 -- # local sub 00:28:50.961 18:14:42 nvmf_dif.fio_dif_digest -- target/dif.sh@45 -- # for sub in "$@" 00:28:50.961 18:14:42 nvmf_dif.fio_dif_digest -- target/dif.sh@46 -- # destroy_subsystem 0 00:28:50.961 18:14:42 nvmf_dif.fio_dif_digest -- target/dif.sh@36 -- # local sub_id=0 00:28:50.961 18:14:42 nvmf_dif.fio_dif_digest -- target/dif.sh@38 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode0 00:28:50.961 18:14:42 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:50.961 18:14:42 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:28:50.961 18:14:42 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:50.961 18:14:42 nvmf_dif.fio_dif_digest -- target/dif.sh@39 -- # rpc_cmd bdev_null_delete bdev_null0 00:28:50.961 18:14:42 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@559 -- # xtrace_disable 00:28:50.961 18:14:42 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:28:50.961 18:14:42 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:28:50.961 00:28:50.961 real 0m11.149s 00:28:50.961 user 0m35.132s 00:28:50.961 sys 0m1.829s 00:28:50.961 18:14:42 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@1124 -- # xtrace_disable 00:28:50.961 18:14:42 nvmf_dif.fio_dif_digest -- common/autotest_common.sh@10 -- # set +x 00:28:50.961 ************************************ 00:28:50.961 END TEST fio_dif_digest 00:28:50.961 ************************************ 00:28:50.961 18:14:42 nvmf_dif -- common/autotest_common.sh@1142 -- # return 0 00:28:50.961 18:14:42 nvmf_dif -- target/dif.sh@146 -- # trap - SIGINT SIGTERM EXIT 00:28:50.961 18:14:42 nvmf_dif -- target/dif.sh@147 -- # nvmftestfini 00:28:50.961 18:14:42 nvmf_dif -- nvmf/common.sh@488 -- # nvmfcleanup 00:28:50.961 18:14:42 nvmf_dif -- nvmf/common.sh@117 -- # sync 00:28:50.961 18:14:42 nvmf_dif -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:28:50.961 18:14:42 nvmf_dif -- nvmf/common.sh@120 -- # set +e 00:28:50.961 18:14:42 nvmf_dif -- nvmf/common.sh@121 -- # for i in {1..20} 00:28:50.961 18:14:42 nvmf_dif -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:28:50.961 rmmod nvme_tcp 00:28:50.961 rmmod nvme_fabrics 00:28:50.961 rmmod nvme_keyring 00:28:50.961 18:14:43 nvmf_dif -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:28:50.961 18:14:43 nvmf_dif -- nvmf/common.sh@124 -- # set -e 00:28:50.961 18:14:43 nvmf_dif -- nvmf/common.sh@125 -- # return 0 00:28:50.961 18:14:43 nvmf_dif -- nvmf/common.sh@489 -- # '[' -n 760791 ']' 00:28:50.961 18:14:43 nvmf_dif -- nvmf/common.sh@490 -- # killprocess 760791 00:28:50.961 18:14:43 nvmf_dif -- common/autotest_common.sh@948 -- # '[' -z 760791 ']' 00:28:50.961 18:14:43 nvmf_dif -- common/autotest_common.sh@952 -- # kill -0 760791 00:28:50.961 18:14:43 nvmf_dif -- common/autotest_common.sh@953 -- # uname 00:28:50.961 18:14:43 nvmf_dif -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:28:50.961 18:14:43 nvmf_dif -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 760791 00:28:50.961 18:14:43 nvmf_dif -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:28:50.961 18:14:43 nvmf_dif -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:28:50.961 18:14:43 nvmf_dif -- common/autotest_common.sh@966 -- # echo 'killing process with pid 760791' 00:28:50.962 killing process with pid 760791 00:28:50.962 18:14:43 nvmf_dif -- common/autotest_common.sh@967 -- # kill 760791 00:28:50.962 18:14:43 nvmf_dif -- common/autotest_common.sh@972 -- # wait 760791 00:28:50.962 18:14:43 nvmf_dif -- nvmf/common.sh@492 -- # '[' iso == iso ']' 00:28:50.962 18:14:43 nvmf_dif -- nvmf/common.sh@493 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:28:51.897 Waiting for block devices as requested 00:28:52.154 0000:5e:00.0 (8086 0a54): vfio-pci -> nvme 00:28:52.154 0000:00:04.7 (8086 2021): vfio-pci -> ioatdma 00:28:52.154 0000:00:04.6 (8086 2021): vfio-pci -> ioatdma 00:28:52.412 0000:00:04.5 (8086 2021): vfio-pci -> ioatdma 00:28:52.412 0000:00:04.4 (8086 2021): vfio-pci -> ioatdma 00:28:52.412 0000:00:04.3 (8086 2021): vfio-pci -> ioatdma 00:28:52.412 0000:00:04.2 (8086 2021): vfio-pci -> ioatdma 00:28:52.671 0000:00:04.1 (8086 2021): vfio-pci -> ioatdma 00:28:52.672 0000:00:04.0 (8086 2021): vfio-pci -> ioatdma 00:28:52.672 0000:80:04.7 (8086 2021): vfio-pci -> ioatdma 00:28:52.672 0000:80:04.6 (8086 2021): vfio-pci -> ioatdma 00:28:52.931 0000:80:04.5 (8086 2021): vfio-pci -> ioatdma 00:28:52.931 0000:80:04.4 (8086 2021): vfio-pci -> ioatdma 00:28:52.931 0000:80:04.3 (8086 2021): vfio-pci -> ioatdma 00:28:52.931 0000:80:04.2 (8086 2021): vfio-pci -> ioatdma 00:28:53.189 0000:80:04.1 (8086 2021): vfio-pci -> ioatdma 00:28:53.189 0000:80:04.0 (8086 2021): vfio-pci -> ioatdma 00:28:53.189 18:14:46 nvmf_dif -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:28:53.189 18:14:46 nvmf_dif -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:28:53.189 18:14:46 nvmf_dif -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:28:53.189 18:14:46 nvmf_dif -- nvmf/common.sh@278 -- # remove_spdk_ns 00:28:53.189 18:14:46 nvmf_dif -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:28:53.189 18:14:46 nvmf_dif -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:28:53.190 18:14:46 nvmf_dif -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:28:55.722 18:14:48 nvmf_dif -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:28:55.722 00:28:55.722 real 1m12.940s 00:28:55.722 user 7m9.765s 00:28:55.722 sys 0m17.949s 00:28:55.722 18:14:48 nvmf_dif -- common/autotest_common.sh@1124 -- # xtrace_disable 00:28:55.722 18:14:48 nvmf_dif -- common/autotest_common.sh@10 -- # set +x 00:28:55.722 ************************************ 00:28:55.722 END TEST nvmf_dif 00:28:55.722 ************************************ 00:28:55.722 18:14:48 -- common/autotest_common.sh@1142 -- # return 0 00:28:55.722 18:14:48 -- spdk/autotest.sh@293 -- # run_test nvmf_abort_qd_sizes /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort_qd_sizes.sh 00:28:55.722 18:14:48 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:28:55.722 18:14:48 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:28:55.722 18:14:48 -- common/autotest_common.sh@10 -- # set +x 00:28:55.722 ************************************ 00:28:55.722 START TEST nvmf_abort_qd_sizes 00:28:55.722 ************************************ 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target/abort_qd_sizes.sh 00:28:55.722 * Looking for test storage... 00:28:55.722 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/target 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@14 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@7 -- # uname -s 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:55.722 18:14:49 nvmf_abort_qd_sizes -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- paths/export.sh@5 -- # export PATH 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@47 -- # : 0 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@51 -- # have_pci_nics=0 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@70 -- # nvmftestinit 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@441 -- # '[' -z tcp ']' 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@446 -- # trap nvmftestfini SIGINT SIGTERM EXIT 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@448 -- # prepare_net_devs 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@410 -- # local -g is_hw=no 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@412 -- # remove_spdk_ns 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@414 -- # [[ phy != virt ]] 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@414 -- # gather_supported_nvmf_pci_devs 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- nvmf/common.sh@285 -- # xtrace_disable 00:28:55.723 18:14:49 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@289 -- # local intel=0x8086 mellanox=0x15b3 pci net_dev 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@291 -- # pci_devs=() 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@291 -- # local -a pci_devs 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@292 -- # pci_net_devs=() 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@292 -- # local -a pci_net_devs 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@293 -- # pci_drivers=() 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@293 -- # local -A pci_drivers 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@295 -- # net_devs=() 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@295 -- # local -ga net_devs 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@296 -- # e810=() 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@296 -- # local -ga e810 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@297 -- # x722=() 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@297 -- # local -ga x722 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@298 -- # mlx=() 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@298 -- # local -ga mlx 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@301 -- # e810+=(${pci_bus_cache["$intel:0x1592"]}) 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@302 -- # e810+=(${pci_bus_cache["$intel:0x159b"]}) 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@304 -- # x722+=(${pci_bus_cache["$intel:0x37d2"]}) 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@306 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2dc"]}) 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@308 -- # mlx+=(${pci_bus_cache["$mellanox:0x1021"]}) 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@310 -- # mlx+=(${pci_bus_cache["$mellanox:0xa2d6"]}) 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@312 -- # mlx+=(${pci_bus_cache["$mellanox:0x101d"]}) 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@314 -- # mlx+=(${pci_bus_cache["$mellanox:0x1017"]}) 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@315 -- # mlx+=(${pci_bus_cache["$mellanox:0x1019"]}) 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@317 -- # mlx+=(${pci_bus_cache["$mellanox:0x1015"]}) 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@318 -- # mlx+=(${pci_bus_cache["$mellanox:0x1013"]}) 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@320 -- # pci_devs+=("${e810[@]}") 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@321 -- # [[ tcp == rdma ]] 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@327 -- # [[ e810 == mlx5 ]] 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@329 -- # [[ e810 == e810 ]] 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@330 -- # pci_devs=("${e810[@]}") 00:29:00.993 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@335 -- # (( 2 == 0 )) 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.0 (0x8086 - 0x159b)' 00:29:00.994 Found 0000:86:00.0 (0x8086 - 0x159b) 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@340 -- # for pci in "${pci_devs[@]}" 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@341 -- # echo 'Found 0000:86:00.1 (0x8086 - 0x159b)' 00:29:00.994 Found 0000:86:00.1 (0x8086 - 0x159b) 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@342 -- # [[ ice == unknown ]] 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@346 -- # [[ ice == unbound ]] 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@350 -- # [[ 0x159b == \0\x\1\0\1\7 ]] 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@351 -- # [[ 0x159b == \0\x\1\0\1\9 ]] 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@352 -- # [[ tcp == rdma ]] 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@366 -- # (( 0 > 0 )) 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@372 -- # [[ e810 == e810 ]] 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@372 -- # [[ tcp == rdma ]] 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@390 -- # [[ up == up ]] 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.0: cvl_0_0' 00:29:00.994 Found net devices under 0000:86:00.0: cvl_0_0 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@382 -- # for pci in "${pci_devs[@]}" 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@383 -- # pci_net_devs=("/sys/bus/pci/devices/$pci/net/"*) 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@388 -- # [[ tcp == tcp ]] 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@389 -- # for net_dev in "${!pci_net_devs[@]}" 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@390 -- # [[ up == up ]] 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@394 -- # (( 1 == 0 )) 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@399 -- # pci_net_devs=("${pci_net_devs[@]##*/}") 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@400 -- # echo 'Found net devices under 0000:86:00.1: cvl_0_1' 00:29:00.994 Found net devices under 0000:86:00.1: cvl_0_1 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@401 -- # net_devs+=("${pci_net_devs[@]}") 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@404 -- # (( 2 == 0 )) 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@414 -- # is_hw=yes 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@416 -- # [[ yes == yes ]] 00:29:00.994 18:14:53 nvmf_abort_qd_sizes -- nvmf/common.sh@417 -- # [[ tcp == tcp ]] 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@418 -- # nvmf_tcp_init 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@229 -- # NVMF_INITIATOR_IP=10.0.0.1 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@230 -- # NVMF_FIRST_TARGET_IP=10.0.0.2 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@231 -- # TCP_INTERFACE_LIST=("${net_devs[@]}") 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@234 -- # (( 2 > 1 )) 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@236 -- # NVMF_TARGET_INTERFACE=cvl_0_0 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@237 -- # NVMF_INITIATOR_INTERFACE=cvl_0_1 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@240 -- # NVMF_SECOND_TARGET_IP= 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@242 -- # NVMF_TARGET_NAMESPACE=cvl_0_0_ns_spdk 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@243 -- # NVMF_TARGET_NS_CMD=(ip netns exec "$NVMF_TARGET_NAMESPACE") 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@244 -- # ip -4 addr flush cvl_0_0 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@245 -- # ip -4 addr flush cvl_0_1 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@248 -- # ip netns add cvl_0_0_ns_spdk 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@251 -- # ip link set cvl_0_0 netns cvl_0_0_ns_spdk 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@254 -- # ip addr add 10.0.0.1/24 dev cvl_0_1 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@255 -- # ip netns exec cvl_0_0_ns_spdk ip addr add 10.0.0.2/24 dev cvl_0_0 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@258 -- # ip link set cvl_0_1 up 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@260 -- # ip netns exec cvl_0_0_ns_spdk ip link set cvl_0_0 up 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@261 -- # ip netns exec cvl_0_0_ns_spdk ip link set lo up 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@264 -- # iptables -I INPUT 1 -i cvl_0_1 -p tcp --dport 4420 -j ACCEPT 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@267 -- # ping -c 1 10.0.0.2 00:29:00.994 PING 10.0.0.2 (10.0.0.2) 56(84) bytes of data. 00:29:00.994 64 bytes from 10.0.0.2: icmp_seq=1 ttl=64 time=0.163 ms 00:29:00.994 00:29:00.994 --- 10.0.0.2 ping statistics --- 00:29:00.994 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:29:00.994 rtt min/avg/max/mdev = 0.163/0.163/0.163/0.000 ms 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@268 -- # ip netns exec cvl_0_0_ns_spdk ping -c 1 10.0.0.1 00:29:00.994 PING 10.0.0.1 (10.0.0.1) 56(84) bytes of data. 00:29:00.994 64 bytes from 10.0.0.1: icmp_seq=1 ttl=64 time=0.096 ms 00:29:00.994 00:29:00.994 --- 10.0.0.1 ping statistics --- 00:29:00.994 1 packets transmitted, 1 received, 0% packet loss, time 0ms 00:29:00.994 rtt min/avg/max/mdev = 0.096/0.096/0.096/0.000 ms 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@270 -- # NVMF_APP=("${NVMF_TARGET_NS_CMD[@]}" "${NVMF_APP[@]}") 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@422 -- # return 0 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@450 -- # '[' iso == iso ']' 00:29:00.994 18:14:54 nvmf_abort_qd_sizes -- nvmf/common.sh@451 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:29:02.900 0000:00:04.7 (8086 2021): ioatdma -> vfio-pci 00:29:02.900 0000:00:04.6 (8086 2021): ioatdma -> vfio-pci 00:29:02.900 0000:00:04.5 (8086 2021): ioatdma -> vfio-pci 00:29:02.900 0000:00:04.4 (8086 2021): ioatdma -> vfio-pci 00:29:02.900 0000:00:04.3 (8086 2021): ioatdma -> vfio-pci 00:29:02.900 0000:00:04.2 (8086 2021): ioatdma -> vfio-pci 00:29:02.900 0000:00:04.1 (8086 2021): ioatdma -> vfio-pci 00:29:02.900 0000:00:04.0 (8086 2021): ioatdma -> vfio-pci 00:29:03.158 0000:80:04.7 (8086 2021): ioatdma -> vfio-pci 00:29:03.158 0000:80:04.6 (8086 2021): ioatdma -> vfio-pci 00:29:03.158 0000:80:04.5 (8086 2021): ioatdma -> vfio-pci 00:29:03.158 0000:80:04.4 (8086 2021): ioatdma -> vfio-pci 00:29:03.158 0000:80:04.3 (8086 2021): ioatdma -> vfio-pci 00:29:03.158 0000:80:04.2 (8086 2021): ioatdma -> vfio-pci 00:29:03.158 0000:80:04.1 (8086 2021): ioatdma -> vfio-pci 00:29:03.158 0000:80:04.0 (8086 2021): ioatdma -> vfio-pci 00:29:04.098 0000:5e:00.0 (8086 0a54): nvme -> vfio-pci 00:29:04.098 18:14:57 nvmf_abort_qd_sizes -- nvmf/common.sh@454 -- # NVMF_TRANSPORT_OPTS='-t tcp' 00:29:04.098 18:14:57 nvmf_abort_qd_sizes -- nvmf/common.sh@455 -- # [[ tcp == \r\d\m\a ]] 00:29:04.098 18:14:57 nvmf_abort_qd_sizes -- nvmf/common.sh@464 -- # [[ tcp == \t\c\p ]] 00:29:04.098 18:14:57 nvmf_abort_qd_sizes -- nvmf/common.sh@465 -- # NVMF_TRANSPORT_OPTS='-t tcp -o' 00:29:04.098 18:14:57 nvmf_abort_qd_sizes -- nvmf/common.sh@468 -- # '[' tcp == tcp ']' 00:29:04.098 18:14:57 nvmf_abort_qd_sizes -- nvmf/common.sh@474 -- # modprobe nvme-tcp 00:29:04.098 18:14:57 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@71 -- # nvmfappstart -m 0xf 00:29:04.098 18:14:57 nvmf_abort_qd_sizes -- nvmf/common.sh@479 -- # timing_enter start_nvmf_tgt 00:29:04.098 18:14:57 nvmf_abort_qd_sizes -- common/autotest_common.sh@722 -- # xtrace_disable 00:29:04.098 18:14:57 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:29:04.098 18:14:57 nvmf_abort_qd_sizes -- nvmf/common.sh@481 -- # nvmfpid=777172 00:29:04.098 18:14:57 nvmf_abort_qd_sizes -- nvmf/common.sh@480 -- # ip netns exec cvl_0_0_ns_spdk /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/nvmf_tgt -i 0 -e 0xFFFF -m 0xf 00:29:04.098 18:14:57 nvmf_abort_qd_sizes -- nvmf/common.sh@482 -- # waitforlisten 777172 00:29:04.098 18:14:57 nvmf_abort_qd_sizes -- common/autotest_common.sh@829 -- # '[' -z 777172 ']' 00:29:04.098 18:14:57 nvmf_abort_qd_sizes -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:04.098 18:14:57 nvmf_abort_qd_sizes -- common/autotest_common.sh@834 -- # local max_retries=100 00:29:04.098 18:14:57 nvmf_abort_qd_sizes -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:04.098 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:04.098 18:14:57 nvmf_abort_qd_sizes -- common/autotest_common.sh@838 -- # xtrace_disable 00:29:04.098 18:14:57 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:29:04.098 [2024-07-15 18:14:57.719391] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:29:04.098 [2024-07-15 18:14:57.719430] [ DPDK EAL parameters: nvmf -c 0xf --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk0 --proc-type=auto ] 00:29:04.098 EAL: No free 2048 kB hugepages reported on node 1 00:29:04.098 [2024-07-15 18:14:57.775841] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 4 00:29:04.357 [2024-07-15 18:14:57.858107] app.c: 603:app_setup_trace: *NOTICE*: Tracepoint Group Mask 0xFFFF specified. 00:29:04.357 [2024-07-15 18:14:57.858145] app.c: 604:app_setup_trace: *NOTICE*: Use 'spdk_trace -s nvmf -i 0' to capture a snapshot of events at runtime. 00:29:04.357 [2024-07-15 18:14:57.858152] app.c: 609:app_setup_trace: *NOTICE*: 'spdk_trace' without parameters will also work if this is the only 00:29:04.357 [2024-07-15 18:14:57.858159] app.c: 610:app_setup_trace: *NOTICE*: SPDK application currently running. 00:29:04.357 [2024-07-15 18:14:57.858164] app.c: 611:app_setup_trace: *NOTICE*: Or copy /dev/shm/nvmf_trace.0 for offline analysis/debug. 00:29:04.357 [2024-07-15 18:14:57.858211] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:29:04.357 [2024-07-15 18:14:57.858312] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 2 00:29:04.357 [2024-07-15 18:14:57.858333] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 3 00:29:04.357 [2024-07-15 18:14:57.858335] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- common/autotest_common.sh@862 -- # return 0 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- nvmf/common.sh@483 -- # timing_exit start_nvmf_tgt 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- common/autotest_common.sh@728 -- # xtrace_disable 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- nvmf/common.sh@484 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini' SIGINT SIGTERM EXIT 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@73 -- # trap 'process_shm --id $NVMF_APP_SHM_ID || :; nvmftestfini || :; clean_kernel_target' SIGINT SIGTERM EXIT 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@75 -- # mapfile -t nvmes 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@75 -- # nvme_in_userspace 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- scripts/common.sh@309 -- # local bdf bdfs 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- scripts/common.sh@310 -- # local nvmes 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- scripts/common.sh@312 -- # [[ -n 0000:5e:00.0 ]] 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- scripts/common.sh@313 -- # nvmes=(${pci_bus_cache["0x010802"]}) 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- scripts/common.sh@318 -- # for bdf in "${nvmes[@]}" 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- scripts/common.sh@319 -- # [[ -e /sys/bus/pci/drivers/nvme/0000:5e:00.0 ]] 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- scripts/common.sh@320 -- # uname -s 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- scripts/common.sh@320 -- # [[ Linux == FreeBSD ]] 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- scripts/common.sh@323 -- # bdfs+=("$bdf") 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- scripts/common.sh@325 -- # (( 1 )) 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- scripts/common.sh@326 -- # printf '%s\n' 0000:5e:00.0 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@76 -- # (( 1 > 0 )) 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@78 -- # nvme=0000:5e:00.0 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@80 -- # run_test spdk_target_abort spdk_target 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- common/autotest_common.sh@1105 -- # xtrace_disable 00:29:04.923 18:14:58 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:29:04.923 ************************************ 00:29:04.923 START TEST spdk_target_abort 00:29:04.923 ************************************ 00:29:04.923 18:14:58 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@1123 -- # spdk_target 00:29:04.923 18:14:58 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@43 -- # local name=spdk_target 00:29:04.923 18:14:58 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@45 -- # rpc_cmd bdev_nvme_attach_controller -t pcie -a 0000:5e:00.0 -b spdk_target 00:29:04.923 18:14:58 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@559 -- # xtrace_disable 00:29:04.923 18:14:58 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:29:08.253 spdk_targetn1 00:29:08.253 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:29:08.253 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@47 -- # rpc_cmd nvmf_create_transport -t tcp -o -u 8192 00:29:08.253 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@559 -- # xtrace_disable 00:29:08.253 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:29:08.253 [2024-07-15 18:15:01.460759] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:29:08.253 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:29:08.253 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@48 -- # rpc_cmd nvmf_create_subsystem nqn.2016-06.io.spdk:testnqn -a -s SPDKISFASTANDAWESOME 00:29:08.253 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@559 -- # xtrace_disable 00:29:08.253 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:29:08.253 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:29:08.253 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@49 -- # rpc_cmd nvmf_subsystem_add_ns nqn.2016-06.io.spdk:testnqn spdk_targetn1 00:29:08.253 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@559 -- # xtrace_disable 00:29:08.253 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:29:08.253 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:29:08.253 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@50 -- # rpc_cmd nvmf_subsystem_add_listener nqn.2016-06.io.spdk:testnqn -t tcp -a 10.0.0.2 -s 4420 00:29:08.253 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@559 -- # xtrace_disable 00:29:08.253 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:29:08.253 [2024-07-15 18:15:01.497678] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 10.0.0.2 port 4420 *** 00:29:08.253 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@52 -- # rabort tcp IPv4 10.0.0.2 4420 nqn.2016-06.io.spdk:testnqn 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@17 -- # local trtype=tcp 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@18 -- # local adrfam=IPv4 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@19 -- # local traddr=10.0.0.2 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@20 -- # local trsvcid=4420 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@21 -- # local subnqn=nqn.2016-06.io.spdk:testnqn 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@23 -- # local qds qd 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@24 -- # local target r 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@26 -- # qds=(4 24 64) 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target=trtype:tcp 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4' 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.2' 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420' 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:29:08.254 18:15:01 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 4 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:29:08.254 EAL: No free 2048 kB hugepages reported on node 1 00:29:11.565 Initializing NVMe Controllers 00:29:11.565 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:testnqn 00:29:11.565 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:29:11.565 Initialization complete. Launching workers. 00:29:11.565 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 15602, failed: 0 00:29:11.565 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 1339, failed to submit 14263 00:29:11.565 success 807, unsuccess 532, failed 0 00:29:11.565 18:15:04 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:29:11.565 18:15:04 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 24 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:29:11.565 EAL: No free 2048 kB hugepages reported on node 1 00:29:14.909 Initializing NVMe Controllers 00:29:14.909 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:testnqn 00:29:14.909 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:29:14.909 Initialization complete. Launching workers. 00:29:14.909 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 8714, failed: 0 00:29:14.909 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 1303, failed to submit 7411 00:29:14.909 success 356, unsuccess 947, failed 0 00:29:14.909 18:15:08 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:29:14.909 18:15:08 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 64 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.2 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:29:14.909 EAL: No free 2048 kB hugepages reported on node 1 00:29:18.195 Initializing NVMe Controllers 00:29:18.195 Attached to NVMe over Fabrics controller at 10.0.0.2:4420: nqn.2016-06.io.spdk:testnqn 00:29:18.195 Associating TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:29:18.195 Initialization complete. Launching workers. 00:29:18.195 NS: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 38085, failed: 0 00:29:18.195 CTRLR: TCP (addr:10.0.0.2 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 2876, failed to submit 35209 00:29:18.195 success 598, unsuccess 2278, failed 0 00:29:18.195 18:15:11 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@54 -- # rpc_cmd nvmf_delete_subsystem nqn.2016-06.io.spdk:testnqn 00:29:18.195 18:15:11 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@559 -- # xtrace_disable 00:29:18.195 18:15:11 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:29:18.195 18:15:11 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:29:18.195 18:15:11 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@55 -- # rpc_cmd bdev_nvme_detach_controller spdk_target 00:29:18.195 18:15:11 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@559 -- # xtrace_disable 00:29:18.195 18:15:11 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:29:19.131 18:15:12 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:29:19.131 18:15:12 nvmf_abort_qd_sizes.spdk_target_abort -- target/abort_qd_sizes.sh@61 -- # killprocess 777172 00:29:19.131 18:15:12 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@948 -- # '[' -z 777172 ']' 00:29:19.131 18:15:12 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@952 -- # kill -0 777172 00:29:19.131 18:15:12 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@953 -- # uname 00:29:19.131 18:15:12 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:29:19.131 18:15:12 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 777172 00:29:19.131 18:15:12 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:29:19.131 18:15:12 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:29:19.131 18:15:12 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@966 -- # echo 'killing process with pid 777172' 00:29:19.131 killing process with pid 777172 00:29:19.131 18:15:12 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@967 -- # kill 777172 00:29:19.131 18:15:12 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@972 -- # wait 777172 00:29:19.131 00:29:19.131 real 0m14.183s 00:29:19.131 user 0m56.626s 00:29:19.131 sys 0m2.248s 00:29:19.131 18:15:12 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@1124 -- # xtrace_disable 00:29:19.131 18:15:12 nvmf_abort_qd_sizes.spdk_target_abort -- common/autotest_common.sh@10 -- # set +x 00:29:19.131 ************************************ 00:29:19.131 END TEST spdk_target_abort 00:29:19.131 ************************************ 00:29:19.131 18:15:12 nvmf_abort_qd_sizes -- common/autotest_common.sh@1142 -- # return 0 00:29:19.131 18:15:12 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@81 -- # run_test kernel_target_abort kernel_target 00:29:19.131 18:15:12 nvmf_abort_qd_sizes -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:29:19.131 18:15:12 nvmf_abort_qd_sizes -- common/autotest_common.sh@1105 -- # xtrace_disable 00:29:19.131 18:15:12 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:29:19.391 ************************************ 00:29:19.391 START TEST kernel_target_abort 00:29:19.391 ************************************ 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1123 -- # kernel_target 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@65 -- # get_main_ns_ip 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@741 -- # local ip 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@742 -- # ip_candidates=() 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@742 -- # local -A ip_candidates 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@744 -- # ip_candidates["rdma"]=NVMF_FIRST_TARGET_IP 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@745 -- # ip_candidates["tcp"]=NVMF_INITIATOR_IP 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@747 -- # [[ -z tcp ]] 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@747 -- # [[ -z NVMF_INITIATOR_IP ]] 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@748 -- # ip=NVMF_INITIATOR_IP 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@750 -- # [[ -z 10.0.0.1 ]] 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@755 -- # echo 10.0.0.1 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@65 -- # configure_kernel_target nqn.2016-06.io.spdk:testnqn 10.0.0.1 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@632 -- # local kernel_name=nqn.2016-06.io.spdk:testnqn kernel_target_ip=10.0.0.1 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@634 -- # nvmet=/sys/kernel/config/nvmet 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@635 -- # kernel_subsystem=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@636 -- # kernel_namespace=/sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@637 -- # kernel_port=/sys/kernel/config/nvmet/ports/1 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@639 -- # local block nvme 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@641 -- # [[ ! -e /sys/module/nvmet ]] 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@642 -- # modprobe nvmet 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@645 -- # [[ -e /sys/kernel/config/nvmet ]] 00:29:19.391 18:15:12 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@647 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:29:21.922 Waiting for block devices as requested 00:29:21.922 0000:5e:00.0 (8086 0a54): vfio-pci -> nvme 00:29:21.922 0000:00:04.7 (8086 2021): vfio-pci -> ioatdma 00:29:21.922 0000:00:04.6 (8086 2021): vfio-pci -> ioatdma 00:29:21.922 0000:00:04.5 (8086 2021): vfio-pci -> ioatdma 00:29:21.922 0000:00:04.4 (8086 2021): vfio-pci -> ioatdma 00:29:21.922 0000:00:04.3 (8086 2021): vfio-pci -> ioatdma 00:29:21.922 0000:00:04.2 (8086 2021): vfio-pci -> ioatdma 00:29:22.181 0000:00:04.1 (8086 2021): vfio-pci -> ioatdma 00:29:22.181 0000:00:04.0 (8086 2021): vfio-pci -> ioatdma 00:29:22.181 0000:80:04.7 (8086 2021): vfio-pci -> ioatdma 00:29:22.181 0000:80:04.6 (8086 2021): vfio-pci -> ioatdma 00:29:22.439 0000:80:04.5 (8086 2021): vfio-pci -> ioatdma 00:29:22.439 0000:80:04.4 (8086 2021): vfio-pci -> ioatdma 00:29:22.439 0000:80:04.3 (8086 2021): vfio-pci -> ioatdma 00:29:22.439 0000:80:04.2 (8086 2021): vfio-pci -> ioatdma 00:29:22.700 0000:80:04.1 (8086 2021): vfio-pci -> ioatdma 00:29:22.700 0000:80:04.0 (8086 2021): vfio-pci -> ioatdma 00:29:22.700 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@650 -- # for block in /sys/block/nvme* 00:29:22.700 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@651 -- # [[ -e /sys/block/nvme0n1 ]] 00:29:22.700 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@652 -- # is_block_zoned nvme0n1 00:29:22.700 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1662 -- # local device=nvme0n1 00:29:22.700 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1664 -- # [[ -e /sys/block/nvme0n1/queue/zoned ]] 00:29:22.700 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1665 -- # [[ none != none ]] 00:29:22.700 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@653 -- # block_in_use nvme0n1 00:29:22.700 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@378 -- # local block=nvme0n1 pt 00:29:22.701 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@387 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/spdk-gpt.py nvme0n1 00:29:22.959 No valid GPT data, bailing 00:29:22.959 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@391 -- # blkid -s PTTYPE -o value /dev/nvme0n1 00:29:22.959 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@391 -- # pt= 00:29:22.959 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- scripts/common.sh@392 -- # return 1 00:29:22.959 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@653 -- # nvme=/dev/nvme0n1 00:29:22.959 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@656 -- # [[ -b /dev/nvme0n1 ]] 00:29:22.959 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@658 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:29:22.959 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@659 -- # mkdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:29:22.959 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@660 -- # mkdir /sys/kernel/config/nvmet/ports/1 00:29:22.959 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@665 -- # echo SPDK-nqn.2016-06.io.spdk:testnqn 00:29:22.959 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@667 -- # echo 1 00:29:22.959 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@668 -- # echo /dev/nvme0n1 00:29:22.959 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@669 -- # echo 1 00:29:22.959 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@671 -- # echo 10.0.0.1 00:29:22.959 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@672 -- # echo tcp 00:29:22.959 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@673 -- # echo 4420 00:29:22.959 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@674 -- # echo ipv4 00:29:22.959 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@677 -- # ln -s /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn /sys/kernel/config/nvmet/ports/1/subsystems/ 00:29:22.959 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@680 -- # nvme discover --hostnqn=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 --hostid=80aaeb9f-0274-ea11-906e-0017a4403562 -a 10.0.0.1 -t tcp -s 4420 00:29:22.959 00:29:22.959 Discovery Log Number of Records 2, Generation counter 2 00:29:22.959 =====Discovery Log Entry 0====== 00:29:22.959 trtype: tcp 00:29:22.959 adrfam: ipv4 00:29:22.959 subtype: current discovery subsystem 00:29:22.959 treq: not specified, sq flow control disable supported 00:29:22.959 portid: 1 00:29:22.959 trsvcid: 4420 00:29:22.959 subnqn: nqn.2014-08.org.nvmexpress.discovery 00:29:22.959 traddr: 10.0.0.1 00:29:22.959 eflags: none 00:29:22.959 sectype: none 00:29:22.959 =====Discovery Log Entry 1====== 00:29:22.959 trtype: tcp 00:29:22.959 adrfam: ipv4 00:29:22.959 subtype: nvme subsystem 00:29:22.959 treq: not specified, sq flow control disable supported 00:29:22.959 portid: 1 00:29:22.959 trsvcid: 4420 00:29:22.959 subnqn: nqn.2016-06.io.spdk:testnqn 00:29:22.959 traddr: 10.0.0.1 00:29:22.959 eflags: none 00:29:22.960 sectype: none 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@66 -- # rabort tcp IPv4 10.0.0.1 4420 nqn.2016-06.io.spdk:testnqn 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@17 -- # local trtype=tcp 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@18 -- # local adrfam=IPv4 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@19 -- # local traddr=10.0.0.1 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@20 -- # local trsvcid=4420 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@21 -- # local subnqn=nqn.2016-06.io.spdk:testnqn 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@23 -- # local qds qd 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@24 -- # local target r 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@26 -- # qds=(4 24 64) 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target=trtype:tcp 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4' 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.1' 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420' 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@28 -- # for r in trtype adrfam traddr trsvcid subnqn 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@29 -- # target='trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:29:22.960 18:15:16 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 4 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:29:22.960 EAL: No free 2048 kB hugepages reported on node 1 00:29:26.247 Initializing NVMe Controllers 00:29:26.247 Attached to NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:29:26.247 Associating TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:29:26.247 Initialization complete. Launching workers. 00:29:26.247 NS: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 77569, failed: 0 00:29:26.247 CTRLR: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 77569, failed to submit 0 00:29:26.247 success 0, unsuccess 77569, failed 0 00:29:26.247 18:15:19 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:29:26.247 18:15:19 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 24 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:29:26.247 EAL: No free 2048 kB hugepages reported on node 1 00:29:29.536 Initializing NVMe Controllers 00:29:29.536 Attached to NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:29:29.536 Associating TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:29:29.536 Initialization complete. Launching workers. 00:29:29.536 NS: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 131366, failed: 0 00:29:29.536 CTRLR: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 32870, failed to submit 98496 00:29:29.536 success 0, unsuccess 32870, failed 0 00:29:29.536 18:15:22 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@32 -- # for qd in "${qds[@]}" 00:29:29.536 18:15:22 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@34 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/abort -q 64 -w rw -M 50 -o 4096 -r 'trtype:tcp adrfam:IPv4 traddr:10.0.0.1 trsvcid:4420 subnqn:nqn.2016-06.io.spdk:testnqn' 00:29:29.536 EAL: No free 2048 kB hugepages reported on node 1 00:29:32.823 Initializing NVMe Controllers 00:29:32.823 Attached to NVMe over Fabrics controller at 10.0.0.1:4420: nqn.2016-06.io.spdk:testnqn 00:29:32.823 Associating TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 with lcore 0 00:29:32.823 Initialization complete. Launching workers. 00:29:32.823 NS: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) NSID 1 I/O completed: 126968, failed: 0 00:29:32.823 CTRLR: TCP (addr:10.0.0.1 subnqn:nqn.2016-06.io.spdk:testnqn) abort submitted 31766, failed to submit 95202 00:29:32.823 success 0, unsuccess 31766, failed 0 00:29:32.823 18:15:25 nvmf_abort_qd_sizes.kernel_target_abort -- target/abort_qd_sizes.sh@67 -- # clean_kernel_target 00:29:32.823 18:15:25 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@684 -- # [[ -e /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn ]] 00:29:32.823 18:15:25 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@686 -- # echo 0 00:29:32.823 18:15:25 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@688 -- # rm -f /sys/kernel/config/nvmet/ports/1/subsystems/nqn.2016-06.io.spdk:testnqn 00:29:32.823 18:15:25 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@689 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn/namespaces/1 00:29:32.823 18:15:25 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@690 -- # rmdir /sys/kernel/config/nvmet/ports/1 00:29:32.823 18:15:25 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@691 -- # rmdir /sys/kernel/config/nvmet/subsystems/nqn.2016-06.io.spdk:testnqn 00:29:32.823 18:15:25 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@693 -- # modules=(/sys/module/nvmet/holders/*) 00:29:32.823 18:15:25 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@695 -- # modprobe -r nvmet_tcp nvmet 00:29:32.823 18:15:25 nvmf_abort_qd_sizes.kernel_target_abort -- nvmf/common.sh@698 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh 00:29:34.730 0000:00:04.7 (8086 2021): ioatdma -> vfio-pci 00:29:34.730 0000:00:04.6 (8086 2021): ioatdma -> vfio-pci 00:29:34.988 0000:00:04.5 (8086 2021): ioatdma -> vfio-pci 00:29:34.988 0000:00:04.4 (8086 2021): ioatdma -> vfio-pci 00:29:34.989 0000:00:04.3 (8086 2021): ioatdma -> vfio-pci 00:29:34.989 0000:00:04.2 (8086 2021): ioatdma -> vfio-pci 00:29:34.989 0000:00:04.1 (8086 2021): ioatdma -> vfio-pci 00:29:34.989 0000:00:04.0 (8086 2021): ioatdma -> vfio-pci 00:29:34.989 0000:80:04.7 (8086 2021): ioatdma -> vfio-pci 00:29:34.989 0000:80:04.6 (8086 2021): ioatdma -> vfio-pci 00:29:34.989 0000:80:04.5 (8086 2021): ioatdma -> vfio-pci 00:29:34.989 0000:80:04.4 (8086 2021): ioatdma -> vfio-pci 00:29:34.989 0000:80:04.3 (8086 2021): ioatdma -> vfio-pci 00:29:34.989 0000:80:04.2 (8086 2021): ioatdma -> vfio-pci 00:29:34.989 0000:80:04.1 (8086 2021): ioatdma -> vfio-pci 00:29:34.989 0000:80:04.0 (8086 2021): ioatdma -> vfio-pci 00:29:35.926 0000:5e:00.0 (8086 0a54): nvme -> vfio-pci 00:29:35.926 00:29:35.926 real 0m16.619s 00:29:35.926 user 0m7.896s 00:29:35.926 sys 0m4.788s 00:29:35.926 18:15:29 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@1124 -- # xtrace_disable 00:29:35.926 18:15:29 nvmf_abort_qd_sizes.kernel_target_abort -- common/autotest_common.sh@10 -- # set +x 00:29:35.926 ************************************ 00:29:35.926 END TEST kernel_target_abort 00:29:35.926 ************************************ 00:29:35.926 18:15:29 nvmf_abort_qd_sizes -- common/autotest_common.sh@1142 -- # return 0 00:29:35.926 18:15:29 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@83 -- # trap - SIGINT SIGTERM EXIT 00:29:35.926 18:15:29 nvmf_abort_qd_sizes -- target/abort_qd_sizes.sh@84 -- # nvmftestfini 00:29:35.926 18:15:29 nvmf_abort_qd_sizes -- nvmf/common.sh@488 -- # nvmfcleanup 00:29:35.926 18:15:29 nvmf_abort_qd_sizes -- nvmf/common.sh@117 -- # sync 00:29:35.926 18:15:29 nvmf_abort_qd_sizes -- nvmf/common.sh@119 -- # '[' tcp == tcp ']' 00:29:35.926 18:15:29 nvmf_abort_qd_sizes -- nvmf/common.sh@120 -- # set +e 00:29:35.926 18:15:29 nvmf_abort_qd_sizes -- nvmf/common.sh@121 -- # for i in {1..20} 00:29:35.926 18:15:29 nvmf_abort_qd_sizes -- nvmf/common.sh@122 -- # modprobe -v -r nvme-tcp 00:29:35.926 rmmod nvme_tcp 00:29:35.926 rmmod nvme_fabrics 00:29:35.926 rmmod nvme_keyring 00:29:35.926 18:15:29 nvmf_abort_qd_sizes -- nvmf/common.sh@123 -- # modprobe -v -r nvme-fabrics 00:29:35.926 18:15:29 nvmf_abort_qd_sizes -- nvmf/common.sh@124 -- # set -e 00:29:35.926 18:15:29 nvmf_abort_qd_sizes -- nvmf/common.sh@125 -- # return 0 00:29:35.926 18:15:29 nvmf_abort_qd_sizes -- nvmf/common.sh@489 -- # '[' -n 777172 ']' 00:29:35.926 18:15:29 nvmf_abort_qd_sizes -- nvmf/common.sh@490 -- # killprocess 777172 00:29:35.926 18:15:29 nvmf_abort_qd_sizes -- common/autotest_common.sh@948 -- # '[' -z 777172 ']' 00:29:35.926 18:15:29 nvmf_abort_qd_sizes -- common/autotest_common.sh@952 -- # kill -0 777172 00:29:35.926 /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/common/autotest_common.sh: line 952: kill: (777172) - No such process 00:29:35.926 18:15:29 nvmf_abort_qd_sizes -- common/autotest_common.sh@975 -- # echo 'Process with pid 777172 is not found' 00:29:35.926 Process with pid 777172 is not found 00:29:35.926 18:15:29 nvmf_abort_qd_sizes -- nvmf/common.sh@492 -- # '[' iso == iso ']' 00:29:35.926 18:15:29 nvmf_abort_qd_sizes -- nvmf/common.sh@493 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/setup.sh reset 00:29:38.470 Waiting for block devices as requested 00:29:38.470 0000:5e:00.0 (8086 0a54): vfio-pci -> nvme 00:29:38.470 0000:00:04.7 (8086 2021): vfio-pci -> ioatdma 00:29:38.470 0000:00:04.6 (8086 2021): vfio-pci -> ioatdma 00:29:38.470 0000:00:04.5 (8086 2021): vfio-pci -> ioatdma 00:29:38.470 0000:00:04.4 (8086 2021): vfio-pci -> ioatdma 00:29:38.470 0000:00:04.3 (8086 2021): vfio-pci -> ioatdma 00:29:38.728 0000:00:04.2 (8086 2021): vfio-pci -> ioatdma 00:29:38.728 0000:00:04.1 (8086 2021): vfio-pci -> ioatdma 00:29:38.728 0000:00:04.0 (8086 2021): vfio-pci -> ioatdma 00:29:38.987 0000:80:04.7 (8086 2021): vfio-pci -> ioatdma 00:29:38.987 0000:80:04.6 (8086 2021): vfio-pci -> ioatdma 00:29:38.988 0000:80:04.5 (8086 2021): vfio-pci -> ioatdma 00:29:38.988 0000:80:04.4 (8086 2021): vfio-pci -> ioatdma 00:29:39.290 0000:80:04.3 (8086 2021): vfio-pci -> ioatdma 00:29:39.290 0000:80:04.2 (8086 2021): vfio-pci -> ioatdma 00:29:39.290 0000:80:04.1 (8086 2021): vfio-pci -> ioatdma 00:29:39.290 0000:80:04.0 (8086 2021): vfio-pci -> ioatdma 00:29:39.549 18:15:33 nvmf_abort_qd_sizes -- nvmf/common.sh@495 -- # [[ tcp == \t\c\p ]] 00:29:39.549 18:15:33 nvmf_abort_qd_sizes -- nvmf/common.sh@496 -- # nvmf_tcp_fini 00:29:39.549 18:15:33 nvmf_abort_qd_sizes -- nvmf/common.sh@274 -- # [[ cvl_0_0_ns_spdk == \n\v\m\f\_\t\g\t\_\n\s ]] 00:29:39.549 18:15:33 nvmf_abort_qd_sizes -- nvmf/common.sh@278 -- # remove_spdk_ns 00:29:39.549 18:15:33 nvmf_abort_qd_sizes -- nvmf/common.sh@628 -- # xtrace_disable_per_cmd _remove_spdk_ns 00:29:39.549 18:15:33 nvmf_abort_qd_sizes -- common/autotest_common.sh@22 -- # eval '_remove_spdk_ns 13> /dev/null' 00:29:39.549 18:15:33 nvmf_abort_qd_sizes -- common/autotest_common.sh@22 -- # _remove_spdk_ns 00:29:41.452 18:15:35 nvmf_abort_qd_sizes -- nvmf/common.sh@279 -- # ip -4 addr flush cvl_0_1 00:29:41.452 00:29:41.452 real 0m46.104s 00:29:41.452 user 1m8.147s 00:29:41.452 sys 0m14.532s 00:29:41.452 18:15:35 nvmf_abort_qd_sizes -- common/autotest_common.sh@1124 -- # xtrace_disable 00:29:41.452 18:15:35 nvmf_abort_qd_sizes -- common/autotest_common.sh@10 -- # set +x 00:29:41.452 ************************************ 00:29:41.452 END TEST nvmf_abort_qd_sizes 00:29:41.452 ************************************ 00:29:41.452 18:15:35 -- common/autotest_common.sh@1142 -- # return 0 00:29:41.452 18:15:35 -- spdk/autotest.sh@295 -- # run_test keyring_file /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/file.sh 00:29:41.452 18:15:35 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:29:41.452 18:15:35 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:29:41.452 18:15:35 -- common/autotest_common.sh@10 -- # set +x 00:29:41.712 ************************************ 00:29:41.712 START TEST keyring_file 00:29:41.712 ************************************ 00:29:41.712 18:15:35 keyring_file -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/file.sh 00:29:41.712 * Looking for test storage... 00:29:41.712 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring 00:29:41.712 18:15:35 keyring_file -- keyring/file.sh@11 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/common.sh 00:29:41.712 18:15:35 keyring_file -- keyring/common.sh@4 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@7 -- # uname -s 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:29:41.712 18:15:35 keyring_file -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:29:41.712 18:15:35 keyring_file -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:29:41.712 18:15:35 keyring_file -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:29:41.712 18:15:35 keyring_file -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:41.712 18:15:35 keyring_file -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:41.712 18:15:35 keyring_file -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:41.712 18:15:35 keyring_file -- paths/export.sh@5 -- # export PATH 00:29:41.712 18:15:35 keyring_file -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@47 -- # : 0 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@51 -- # have_pci_nics=0 00:29:41.712 18:15:35 keyring_file -- keyring/common.sh@6 -- # bperfsock=/var/tmp/bperf.sock 00:29:41.712 18:15:35 keyring_file -- keyring/file.sh@13 -- # subnqn=nqn.2016-06.io.spdk:cnode0 00:29:41.712 18:15:35 keyring_file -- keyring/file.sh@14 -- # hostnqn=nqn.2016-06.io.spdk:host0 00:29:41.712 18:15:35 keyring_file -- keyring/file.sh@15 -- # key0=00112233445566778899aabbccddeeff 00:29:41.712 18:15:35 keyring_file -- keyring/file.sh@16 -- # key1=112233445566778899aabbccddeeff00 00:29:41.712 18:15:35 keyring_file -- keyring/file.sh@24 -- # trap cleanup EXIT 00:29:41.712 18:15:35 keyring_file -- keyring/file.sh@26 -- # prep_key key0 00112233445566778899aabbccddeeff 0 00:29:41.712 18:15:35 keyring_file -- keyring/common.sh@15 -- # local name key digest path 00:29:41.712 18:15:35 keyring_file -- keyring/common.sh@17 -- # name=key0 00:29:41.712 18:15:35 keyring_file -- keyring/common.sh@17 -- # key=00112233445566778899aabbccddeeff 00:29:41.712 18:15:35 keyring_file -- keyring/common.sh@17 -- # digest=0 00:29:41.712 18:15:35 keyring_file -- keyring/common.sh@18 -- # mktemp 00:29:41.712 18:15:35 keyring_file -- keyring/common.sh@18 -- # path=/tmp/tmp.UsIn14501G 00:29:41.712 18:15:35 keyring_file -- keyring/common.sh@20 -- # format_interchange_psk 00112233445566778899aabbccddeeff 0 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@715 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 0 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@702 -- # local prefix key digest 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@704 -- # prefix=NVMeTLSkey-1 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@704 -- # key=00112233445566778899aabbccddeeff 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@704 -- # digest=0 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@705 -- # python - 00:29:41.712 18:15:35 keyring_file -- keyring/common.sh@21 -- # chmod 0600 /tmp/tmp.UsIn14501G 00:29:41.712 18:15:35 keyring_file -- keyring/common.sh@23 -- # echo /tmp/tmp.UsIn14501G 00:29:41.712 18:15:35 keyring_file -- keyring/file.sh@26 -- # key0path=/tmp/tmp.UsIn14501G 00:29:41.712 18:15:35 keyring_file -- keyring/file.sh@27 -- # prep_key key1 112233445566778899aabbccddeeff00 0 00:29:41.712 18:15:35 keyring_file -- keyring/common.sh@15 -- # local name key digest path 00:29:41.712 18:15:35 keyring_file -- keyring/common.sh@17 -- # name=key1 00:29:41.712 18:15:35 keyring_file -- keyring/common.sh@17 -- # key=112233445566778899aabbccddeeff00 00:29:41.712 18:15:35 keyring_file -- keyring/common.sh@17 -- # digest=0 00:29:41.712 18:15:35 keyring_file -- keyring/common.sh@18 -- # mktemp 00:29:41.712 18:15:35 keyring_file -- keyring/common.sh@18 -- # path=/tmp/tmp.REXsPcHwDh 00:29:41.712 18:15:35 keyring_file -- keyring/common.sh@20 -- # format_interchange_psk 112233445566778899aabbccddeeff00 0 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@715 -- # format_key NVMeTLSkey-1 112233445566778899aabbccddeeff00 0 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@702 -- # local prefix key digest 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@704 -- # prefix=NVMeTLSkey-1 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@704 -- # key=112233445566778899aabbccddeeff00 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@704 -- # digest=0 00:29:41.712 18:15:35 keyring_file -- nvmf/common.sh@705 -- # python - 00:29:41.712 18:15:35 keyring_file -- keyring/common.sh@21 -- # chmod 0600 /tmp/tmp.REXsPcHwDh 00:29:41.712 18:15:35 keyring_file -- keyring/common.sh@23 -- # echo /tmp/tmp.REXsPcHwDh 00:29:41.712 18:15:35 keyring_file -- keyring/file.sh@27 -- # key1path=/tmp/tmp.REXsPcHwDh 00:29:41.712 18:15:35 keyring_file -- keyring/file.sh@30 -- # tgtpid=786331 00:29:41.712 18:15:35 keyring_file -- keyring/file.sh@29 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:29:41.712 18:15:35 keyring_file -- keyring/file.sh@32 -- # waitforlisten 786331 00:29:41.712 18:15:35 keyring_file -- common/autotest_common.sh@829 -- # '[' -z 786331 ']' 00:29:41.712 18:15:35 keyring_file -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:41.712 18:15:35 keyring_file -- common/autotest_common.sh@834 -- # local max_retries=100 00:29:41.712 18:15:35 keyring_file -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:41.712 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:41.712 18:15:35 keyring_file -- common/autotest_common.sh@838 -- # xtrace_disable 00:29:41.712 18:15:35 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:29:41.972 [2024-07-15 18:15:35.459064] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:29:41.972 [2024-07-15 18:15:35.459115] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid786331 ] 00:29:41.972 EAL: No free 2048 kB hugepages reported on node 1 00:29:41.972 [2024-07-15 18:15:35.512735] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:41.972 [2024-07-15 18:15:35.593126] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:29:42.541 18:15:36 keyring_file -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:29:42.541 18:15:36 keyring_file -- common/autotest_common.sh@862 -- # return 0 00:29:42.541 18:15:36 keyring_file -- keyring/file.sh@33 -- # rpc_cmd 00:29:42.541 18:15:36 keyring_file -- common/autotest_common.sh@559 -- # xtrace_disable 00:29:42.541 18:15:36 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:29:42.541 [2024-07-15 18:15:36.265550] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:29:42.800 null0 00:29:42.800 [2024-07-15 18:15:36.297591] tcp.c: 966:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:29:42.800 [2024-07-15 18:15:36.297845] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:29:42.800 [2024-07-15 18:15:36.305607] tcp.c:3710:nvmf_tcp_subsystem_add_host: *WARNING*: nvmf_tcp_psk_path: deprecated feature PSK path to be removed in v24.09 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:29:42.800 18:15:36 keyring_file -- keyring/file.sh@43 -- # NOT rpc_cmd nvmf_subsystem_add_listener -t tcp -a 127.0.0.1 -s 4420 nqn.2016-06.io.spdk:cnode0 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@648 -- # local es=0 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@650 -- # valid_exec_arg rpc_cmd nvmf_subsystem_add_listener -t tcp -a 127.0.0.1 -s 4420 nqn.2016-06.io.spdk:cnode0 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@636 -- # local arg=rpc_cmd 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@640 -- # type -t rpc_cmd 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@651 -- # rpc_cmd nvmf_subsystem_add_listener -t tcp -a 127.0.0.1 -s 4420 nqn.2016-06.io.spdk:cnode0 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@559 -- # xtrace_disable 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:29:42.800 [2024-07-15 18:15:36.317640] nvmf_rpc.c: 783:nvmf_rpc_listen_paused: *ERROR*: Listener already exists 00:29:42.800 request: 00:29:42.800 { 00:29:42.800 "nqn": "nqn.2016-06.io.spdk:cnode0", 00:29:42.800 "secure_channel": false, 00:29:42.800 "listen_address": { 00:29:42.800 "trtype": "tcp", 00:29:42.800 "traddr": "127.0.0.1", 00:29:42.800 "trsvcid": "4420" 00:29:42.800 }, 00:29:42.800 "method": "nvmf_subsystem_add_listener", 00:29:42.800 "req_id": 1 00:29:42.800 } 00:29:42.800 Got JSON-RPC error response 00:29:42.800 response: 00:29:42.800 { 00:29:42.800 "code": -32602, 00:29:42.800 "message": "Invalid parameters" 00:29:42.800 } 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@587 -- # [[ 1 == 0 ]] 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@651 -- # es=1 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:29:42.800 18:15:36 keyring_file -- keyring/file.sh@46 -- # bperfpid=786467 00:29:42.800 18:15:36 keyring_file -- keyring/file.sh@45 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -q 128 -o 4k -w randrw -M 50 -t 1 -m 2 -r /var/tmp/bperf.sock -z 00:29:42.800 18:15:36 keyring_file -- keyring/file.sh@48 -- # waitforlisten 786467 /var/tmp/bperf.sock 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@829 -- # '[' -z 786467 ']' 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bperf.sock 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@834 -- # local max_retries=100 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:29:42.800 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@838 -- # xtrace_disable 00:29:42.800 18:15:36 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:29:42.800 [2024-07-15 18:15:36.366951] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:29:42.800 [2024-07-15 18:15:36.366990] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid786467 ] 00:29:42.800 EAL: No free 2048 kB hugepages reported on node 1 00:29:42.800 [2024-07-15 18:15:36.419592] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:42.800 [2024-07-15 18:15:36.497997] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:29:43.736 18:15:37 keyring_file -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:29:43.736 18:15:37 keyring_file -- common/autotest_common.sh@862 -- # return 0 00:29:43.736 18:15:37 keyring_file -- keyring/file.sh@49 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.UsIn14501G 00:29:43.736 18:15:37 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.UsIn14501G 00:29:43.736 18:15:37 keyring_file -- keyring/file.sh@50 -- # bperf_cmd keyring_file_add_key key1 /tmp/tmp.REXsPcHwDh 00:29:43.736 18:15:37 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key1 /tmp/tmp.REXsPcHwDh 00:29:43.994 18:15:37 keyring_file -- keyring/file.sh@51 -- # get_key key0 00:29:43.994 18:15:37 keyring_file -- keyring/file.sh@51 -- # jq -r .path 00:29:43.994 18:15:37 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:29:43.994 18:15:37 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:29:43.994 18:15:37 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:43.994 18:15:37 keyring_file -- keyring/file.sh@51 -- # [[ /tmp/tmp.UsIn14501G == \/\t\m\p\/\t\m\p\.\U\s\I\n\1\4\5\0\1\G ]] 00:29:43.994 18:15:37 keyring_file -- keyring/file.sh@52 -- # get_key key1 00:29:43.994 18:15:37 keyring_file -- keyring/file.sh@52 -- # jq -r .path 00:29:43.994 18:15:37 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:29:43.994 18:15:37 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:29:43.994 18:15:37 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:44.252 18:15:37 keyring_file -- keyring/file.sh@52 -- # [[ /tmp/tmp.REXsPcHwDh == \/\t\m\p\/\t\m\p\.\R\E\X\s\P\c\H\w\D\h ]] 00:29:44.252 18:15:37 keyring_file -- keyring/file.sh@53 -- # get_refcnt key0 00:29:44.252 18:15:37 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:29:44.252 18:15:37 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:29:44.252 18:15:37 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:29:44.252 18:15:37 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:44.252 18:15:37 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:29:44.511 18:15:38 keyring_file -- keyring/file.sh@53 -- # (( 1 == 1 )) 00:29:44.511 18:15:38 keyring_file -- keyring/file.sh@54 -- # get_refcnt key1 00:29:44.511 18:15:38 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:29:44.511 18:15:38 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:29:44.511 18:15:38 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:29:44.511 18:15:38 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:29:44.511 18:15:38 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:44.511 18:15:38 keyring_file -- keyring/file.sh@54 -- # (( 1 == 1 )) 00:29:44.511 18:15:38 keyring_file -- keyring/file.sh@57 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:29:44.511 18:15:38 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:29:44.770 [2024-07-15 18:15:38.374952] bdev_nvme_rpc.c: 517:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:29:44.770 nvme0n1 00:29:44.770 18:15:38 keyring_file -- keyring/file.sh@59 -- # get_refcnt key0 00:29:44.770 18:15:38 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:29:44.770 18:15:38 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:29:44.770 18:15:38 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:29:44.770 18:15:38 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:29:44.770 18:15:38 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:45.029 18:15:38 keyring_file -- keyring/file.sh@59 -- # (( 2 == 2 )) 00:29:45.029 18:15:38 keyring_file -- keyring/file.sh@60 -- # get_refcnt key1 00:29:45.029 18:15:38 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:29:45.029 18:15:38 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:29:45.029 18:15:38 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:29:45.029 18:15:38 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:45.029 18:15:38 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:29:45.288 18:15:38 keyring_file -- keyring/file.sh@60 -- # (( 1 == 1 )) 00:29:45.288 18:15:38 keyring_file -- keyring/file.sh@62 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:29:45.288 Running I/O for 1 seconds... 00:29:46.225 00:29:46.225 Latency(us) 00:29:46.225 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:46.225 Job: nvme0n1 (Core Mask 0x2, workload: randrw, percentage: 50, depth: 128, IO size: 4096) 00:29:46.225 nvme0n1 : 1.01 13669.22 53.40 0.00 0.00 9319.60 7864.32 19603.81 00:29:46.225 =================================================================================================================== 00:29:46.225 Total : 13669.22 53.40 0.00 0.00 9319.60 7864.32 19603.81 00:29:46.225 0 00:29:46.225 18:15:39 keyring_file -- keyring/file.sh@64 -- # bperf_cmd bdev_nvme_detach_controller nvme0 00:29:46.225 18:15:39 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_detach_controller nvme0 00:29:46.484 18:15:40 keyring_file -- keyring/file.sh@65 -- # get_refcnt key0 00:29:46.484 18:15:40 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:29:46.484 18:15:40 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:29:46.484 18:15:40 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:29:46.484 18:15:40 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:29:46.484 18:15:40 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:46.743 18:15:40 keyring_file -- keyring/file.sh@65 -- # (( 1 == 1 )) 00:29:46.743 18:15:40 keyring_file -- keyring/file.sh@66 -- # get_refcnt key1 00:29:46.743 18:15:40 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:29:46.743 18:15:40 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:29:46.743 18:15:40 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:29:46.743 18:15:40 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:29:46.743 18:15:40 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:46.743 18:15:40 keyring_file -- keyring/file.sh@66 -- # (( 1 == 1 )) 00:29:46.743 18:15:40 keyring_file -- keyring/file.sh@69 -- # NOT bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:29:46.743 18:15:40 keyring_file -- common/autotest_common.sh@648 -- # local es=0 00:29:47.002 18:15:40 keyring_file -- common/autotest_common.sh@650 -- # valid_exec_arg bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:29:47.002 18:15:40 keyring_file -- common/autotest_common.sh@636 -- # local arg=bperf_cmd 00:29:47.002 18:15:40 keyring_file -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:29:47.002 18:15:40 keyring_file -- common/autotest_common.sh@640 -- # type -t bperf_cmd 00:29:47.002 18:15:40 keyring_file -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:29:47.002 18:15:40 keyring_file -- common/autotest_common.sh@651 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:29:47.002 18:15:40 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key1 00:29:47.002 [2024-07-15 18:15:40.630681] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 428:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:29:47.002 [2024-07-15 18:15:40.630732] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x677770 (107): Transport endpoint is not connected 00:29:47.002 [2024-07-15 18:15:40.631727] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x677770 (9): Bad file descriptor 00:29:47.002 [2024-07-15 18:15:40.632728] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:29:47.002 [2024-07-15 18:15:40.632737] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 127.0.0.1 00:29:47.002 [2024-07-15 18:15:40.632744] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:29:47.002 request: 00:29:47.002 { 00:29:47.002 "name": "nvme0", 00:29:47.002 "trtype": "tcp", 00:29:47.002 "traddr": "127.0.0.1", 00:29:47.002 "adrfam": "ipv4", 00:29:47.002 "trsvcid": "4420", 00:29:47.002 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:29:47.002 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:29:47.002 "prchk_reftag": false, 00:29:47.002 "prchk_guard": false, 00:29:47.002 "hdgst": false, 00:29:47.002 "ddgst": false, 00:29:47.002 "psk": "key1", 00:29:47.002 "method": "bdev_nvme_attach_controller", 00:29:47.002 "req_id": 1 00:29:47.002 } 00:29:47.002 Got JSON-RPC error response 00:29:47.002 response: 00:29:47.002 { 00:29:47.002 "code": -5, 00:29:47.002 "message": "Input/output error" 00:29:47.002 } 00:29:47.002 18:15:40 keyring_file -- common/autotest_common.sh@651 -- # es=1 00:29:47.002 18:15:40 keyring_file -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:29:47.002 18:15:40 keyring_file -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:29:47.002 18:15:40 keyring_file -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:29:47.002 18:15:40 keyring_file -- keyring/file.sh@71 -- # get_refcnt key0 00:29:47.002 18:15:40 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:29:47.002 18:15:40 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:29:47.002 18:15:40 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:29:47.002 18:15:40 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:29:47.002 18:15:40 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:47.261 18:15:40 keyring_file -- keyring/file.sh@71 -- # (( 1 == 1 )) 00:29:47.261 18:15:40 keyring_file -- keyring/file.sh@72 -- # get_refcnt key1 00:29:47.261 18:15:40 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:29:47.261 18:15:40 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:29:47.261 18:15:40 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:29:47.261 18:15:40 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:29:47.261 18:15:40 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:47.519 18:15:41 keyring_file -- keyring/file.sh@72 -- # (( 1 == 1 )) 00:29:47.520 18:15:41 keyring_file -- keyring/file.sh@75 -- # bperf_cmd keyring_file_remove_key key0 00:29:47.520 18:15:41 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key0 00:29:47.520 18:15:41 keyring_file -- keyring/file.sh@76 -- # bperf_cmd keyring_file_remove_key key1 00:29:47.520 18:15:41 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key1 00:29:47.778 18:15:41 keyring_file -- keyring/file.sh@77 -- # bperf_cmd keyring_get_keys 00:29:47.778 18:15:41 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:47.778 18:15:41 keyring_file -- keyring/file.sh@77 -- # jq length 00:29:48.037 18:15:41 keyring_file -- keyring/file.sh@77 -- # (( 0 == 0 )) 00:29:48.037 18:15:41 keyring_file -- keyring/file.sh@80 -- # chmod 0660 /tmp/tmp.UsIn14501G 00:29:48.037 18:15:41 keyring_file -- keyring/file.sh@81 -- # NOT bperf_cmd keyring_file_add_key key0 /tmp/tmp.UsIn14501G 00:29:48.037 18:15:41 keyring_file -- common/autotest_common.sh@648 -- # local es=0 00:29:48.037 18:15:41 keyring_file -- common/autotest_common.sh@650 -- # valid_exec_arg bperf_cmd keyring_file_add_key key0 /tmp/tmp.UsIn14501G 00:29:48.037 18:15:41 keyring_file -- common/autotest_common.sh@636 -- # local arg=bperf_cmd 00:29:48.037 18:15:41 keyring_file -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:29:48.037 18:15:41 keyring_file -- common/autotest_common.sh@640 -- # type -t bperf_cmd 00:29:48.037 18:15:41 keyring_file -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:29:48.037 18:15:41 keyring_file -- common/autotest_common.sh@651 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.UsIn14501G 00:29:48.037 18:15:41 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.UsIn14501G 00:29:48.037 [2024-07-15 18:15:41.664858] keyring.c: 34:keyring_file_check_path: *ERROR*: Invalid permissions for key file '/tmp/tmp.UsIn14501G': 0100660 00:29:48.037 [2024-07-15 18:15:41.664883] keyring.c: 126:spdk_keyring_add_key: *ERROR*: Failed to add key 'key0' to the keyring 00:29:48.037 request: 00:29:48.037 { 00:29:48.037 "name": "key0", 00:29:48.037 "path": "/tmp/tmp.UsIn14501G", 00:29:48.037 "method": "keyring_file_add_key", 00:29:48.037 "req_id": 1 00:29:48.037 } 00:29:48.037 Got JSON-RPC error response 00:29:48.037 response: 00:29:48.037 { 00:29:48.037 "code": -1, 00:29:48.037 "message": "Operation not permitted" 00:29:48.037 } 00:29:48.037 18:15:41 keyring_file -- common/autotest_common.sh@651 -- # es=1 00:29:48.037 18:15:41 keyring_file -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:29:48.037 18:15:41 keyring_file -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:29:48.037 18:15:41 keyring_file -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:29:48.037 18:15:41 keyring_file -- keyring/file.sh@84 -- # chmod 0600 /tmp/tmp.UsIn14501G 00:29:48.037 18:15:41 keyring_file -- keyring/file.sh@85 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.UsIn14501G 00:29:48.037 18:15:41 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.UsIn14501G 00:29:48.296 18:15:41 keyring_file -- keyring/file.sh@86 -- # rm -f /tmp/tmp.UsIn14501G 00:29:48.296 18:15:41 keyring_file -- keyring/file.sh@88 -- # get_refcnt key0 00:29:48.296 18:15:41 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:29:48.296 18:15:41 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:29:48.296 18:15:41 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:29:48.296 18:15:41 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:48.296 18:15:41 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:29:48.555 18:15:42 keyring_file -- keyring/file.sh@88 -- # (( 1 == 1 )) 00:29:48.555 18:15:42 keyring_file -- keyring/file.sh@90 -- # NOT bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:29:48.555 18:15:42 keyring_file -- common/autotest_common.sh@648 -- # local es=0 00:29:48.555 18:15:42 keyring_file -- common/autotest_common.sh@650 -- # valid_exec_arg bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:29:48.555 18:15:42 keyring_file -- common/autotest_common.sh@636 -- # local arg=bperf_cmd 00:29:48.555 18:15:42 keyring_file -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:29:48.555 18:15:42 keyring_file -- common/autotest_common.sh@640 -- # type -t bperf_cmd 00:29:48.555 18:15:42 keyring_file -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:29:48.555 18:15:42 keyring_file -- common/autotest_common.sh@651 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:29:48.555 18:15:42 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:29:48.555 [2024-07-15 18:15:42.190252] keyring.c: 29:keyring_file_check_path: *ERROR*: Could not stat key file '/tmp/tmp.UsIn14501G': No such file or directory 00:29:48.555 [2024-07-15 18:15:42.190270] nvme_tcp.c:2582:nvme_tcp_generate_tls_credentials: *ERROR*: Failed to obtain key 'key0': No such file or directory 00:29:48.555 [2024-07-15 18:15:42.190291] nvme.c: 683:nvme_ctrlr_probe: *ERROR*: Failed to construct NVMe controller for SSD: 127.0.0.1 00:29:48.555 [2024-07-15 18:15:42.190296] nvme.c: 830:nvme_probe_internal: *ERROR*: NVMe ctrlr scan failed 00:29:48.555 [2024-07-15 18:15:42.190302] bdev_nvme.c:6268:bdev_nvme_create: *ERROR*: No controller was found with provided trid (traddr: 127.0.0.1) 00:29:48.555 request: 00:29:48.555 { 00:29:48.555 "name": "nvme0", 00:29:48.555 "trtype": "tcp", 00:29:48.555 "traddr": "127.0.0.1", 00:29:48.555 "adrfam": "ipv4", 00:29:48.555 "trsvcid": "4420", 00:29:48.555 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:29:48.555 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:29:48.555 "prchk_reftag": false, 00:29:48.555 "prchk_guard": false, 00:29:48.555 "hdgst": false, 00:29:48.555 "ddgst": false, 00:29:48.555 "psk": "key0", 00:29:48.555 "method": "bdev_nvme_attach_controller", 00:29:48.555 "req_id": 1 00:29:48.555 } 00:29:48.555 Got JSON-RPC error response 00:29:48.555 response: 00:29:48.555 { 00:29:48.555 "code": -19, 00:29:48.555 "message": "No such device" 00:29:48.555 } 00:29:48.555 18:15:42 keyring_file -- common/autotest_common.sh@651 -- # es=1 00:29:48.555 18:15:42 keyring_file -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:29:48.555 18:15:42 keyring_file -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:29:48.555 18:15:42 keyring_file -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:29:48.555 18:15:42 keyring_file -- keyring/file.sh@92 -- # bperf_cmd keyring_file_remove_key key0 00:29:48.555 18:15:42 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key0 00:29:48.814 18:15:42 keyring_file -- keyring/file.sh@95 -- # prep_key key0 00112233445566778899aabbccddeeff 0 00:29:48.814 18:15:42 keyring_file -- keyring/common.sh@15 -- # local name key digest path 00:29:48.814 18:15:42 keyring_file -- keyring/common.sh@17 -- # name=key0 00:29:48.814 18:15:42 keyring_file -- keyring/common.sh@17 -- # key=00112233445566778899aabbccddeeff 00:29:48.814 18:15:42 keyring_file -- keyring/common.sh@17 -- # digest=0 00:29:48.814 18:15:42 keyring_file -- keyring/common.sh@18 -- # mktemp 00:29:48.814 18:15:42 keyring_file -- keyring/common.sh@18 -- # path=/tmp/tmp.xPkaUTaC9A 00:29:48.814 18:15:42 keyring_file -- keyring/common.sh@20 -- # format_interchange_psk 00112233445566778899aabbccddeeff 0 00:29:48.814 18:15:42 keyring_file -- nvmf/common.sh@715 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 0 00:29:48.814 18:15:42 keyring_file -- nvmf/common.sh@702 -- # local prefix key digest 00:29:48.814 18:15:42 keyring_file -- nvmf/common.sh@704 -- # prefix=NVMeTLSkey-1 00:29:48.814 18:15:42 keyring_file -- nvmf/common.sh@704 -- # key=00112233445566778899aabbccddeeff 00:29:48.814 18:15:42 keyring_file -- nvmf/common.sh@704 -- # digest=0 00:29:48.814 18:15:42 keyring_file -- nvmf/common.sh@705 -- # python - 00:29:48.814 18:15:42 keyring_file -- keyring/common.sh@21 -- # chmod 0600 /tmp/tmp.xPkaUTaC9A 00:29:48.814 18:15:42 keyring_file -- keyring/common.sh@23 -- # echo /tmp/tmp.xPkaUTaC9A 00:29:48.814 18:15:42 keyring_file -- keyring/file.sh@95 -- # key0path=/tmp/tmp.xPkaUTaC9A 00:29:48.814 18:15:42 keyring_file -- keyring/file.sh@96 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.xPkaUTaC9A 00:29:48.814 18:15:42 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.xPkaUTaC9A 00:29:49.073 18:15:42 keyring_file -- keyring/file.sh@97 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:29:49.073 18:15:42 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:29:49.332 nvme0n1 00:29:49.332 18:15:42 keyring_file -- keyring/file.sh@99 -- # get_refcnt key0 00:29:49.332 18:15:42 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:29:49.332 18:15:42 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:29:49.332 18:15:42 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:29:49.332 18:15:42 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:29:49.332 18:15:42 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:49.332 18:15:43 keyring_file -- keyring/file.sh@99 -- # (( 2 == 2 )) 00:29:49.332 18:15:43 keyring_file -- keyring/file.sh@100 -- # bperf_cmd keyring_file_remove_key key0 00:29:49.332 18:15:43 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_remove_key key0 00:29:49.590 18:15:43 keyring_file -- keyring/file.sh@101 -- # get_key key0 00:29:49.590 18:15:43 keyring_file -- keyring/file.sh@101 -- # jq -r .removed 00:29:49.590 18:15:43 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:29:49.590 18:15:43 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:29:49.590 18:15:43 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:49.848 18:15:43 keyring_file -- keyring/file.sh@101 -- # [[ true == \t\r\u\e ]] 00:29:49.848 18:15:43 keyring_file -- keyring/file.sh@102 -- # get_refcnt key0 00:29:49.848 18:15:43 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:29:49.848 18:15:43 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:29:49.848 18:15:43 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:29:49.848 18:15:43 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:29:49.848 18:15:43 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:50.107 18:15:43 keyring_file -- keyring/file.sh@102 -- # (( 1 == 1 )) 00:29:50.107 18:15:43 keyring_file -- keyring/file.sh@103 -- # bperf_cmd bdev_nvme_detach_controller nvme0 00:29:50.107 18:15:43 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_detach_controller nvme0 00:29:50.107 18:15:43 keyring_file -- keyring/file.sh@104 -- # bperf_cmd keyring_get_keys 00:29:50.107 18:15:43 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:50.107 18:15:43 keyring_file -- keyring/file.sh@104 -- # jq length 00:29:50.366 18:15:43 keyring_file -- keyring/file.sh@104 -- # (( 0 == 0 )) 00:29:50.366 18:15:43 keyring_file -- keyring/file.sh@107 -- # bperf_cmd keyring_file_add_key key0 /tmp/tmp.xPkaUTaC9A 00:29:50.366 18:15:43 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key0 /tmp/tmp.xPkaUTaC9A 00:29:50.625 18:15:44 keyring_file -- keyring/file.sh@108 -- # bperf_cmd keyring_file_add_key key1 /tmp/tmp.REXsPcHwDh 00:29:50.625 18:15:44 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_file_add_key key1 /tmp/tmp.REXsPcHwDh 00:29:50.625 18:15:44 keyring_file -- keyring/file.sh@109 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:29:50.625 18:15:44 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk key0 00:29:50.883 nvme0n1 00:29:50.883 18:15:44 keyring_file -- keyring/file.sh@112 -- # bperf_cmd save_config 00:29:50.883 18:15:44 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock save_config 00:29:51.142 18:15:44 keyring_file -- keyring/file.sh@112 -- # config='{ 00:29:51.142 "subsystems": [ 00:29:51.142 { 00:29:51.142 "subsystem": "keyring", 00:29:51.142 "config": [ 00:29:51.142 { 00:29:51.142 "method": "keyring_file_add_key", 00:29:51.142 "params": { 00:29:51.142 "name": "key0", 00:29:51.142 "path": "/tmp/tmp.xPkaUTaC9A" 00:29:51.142 } 00:29:51.142 }, 00:29:51.142 { 00:29:51.142 "method": "keyring_file_add_key", 00:29:51.142 "params": { 00:29:51.142 "name": "key1", 00:29:51.142 "path": "/tmp/tmp.REXsPcHwDh" 00:29:51.142 } 00:29:51.142 } 00:29:51.142 ] 00:29:51.142 }, 00:29:51.142 { 00:29:51.142 "subsystem": "iobuf", 00:29:51.142 "config": [ 00:29:51.142 { 00:29:51.142 "method": "iobuf_set_options", 00:29:51.142 "params": { 00:29:51.142 "small_pool_count": 8192, 00:29:51.142 "large_pool_count": 1024, 00:29:51.142 "small_bufsize": 8192, 00:29:51.142 "large_bufsize": 135168 00:29:51.142 } 00:29:51.142 } 00:29:51.142 ] 00:29:51.142 }, 00:29:51.142 { 00:29:51.142 "subsystem": "sock", 00:29:51.142 "config": [ 00:29:51.142 { 00:29:51.142 "method": "sock_set_default_impl", 00:29:51.142 "params": { 00:29:51.142 "impl_name": "posix" 00:29:51.142 } 00:29:51.142 }, 00:29:51.142 { 00:29:51.142 "method": "sock_impl_set_options", 00:29:51.142 "params": { 00:29:51.142 "impl_name": "ssl", 00:29:51.142 "recv_buf_size": 4096, 00:29:51.142 "send_buf_size": 4096, 00:29:51.142 "enable_recv_pipe": true, 00:29:51.142 "enable_quickack": false, 00:29:51.142 "enable_placement_id": 0, 00:29:51.142 "enable_zerocopy_send_server": true, 00:29:51.142 "enable_zerocopy_send_client": false, 00:29:51.142 "zerocopy_threshold": 0, 00:29:51.142 "tls_version": 0, 00:29:51.142 "enable_ktls": false 00:29:51.142 } 00:29:51.142 }, 00:29:51.142 { 00:29:51.142 "method": "sock_impl_set_options", 00:29:51.142 "params": { 00:29:51.142 "impl_name": "posix", 00:29:51.142 "recv_buf_size": 2097152, 00:29:51.142 "send_buf_size": 2097152, 00:29:51.142 "enable_recv_pipe": true, 00:29:51.142 "enable_quickack": false, 00:29:51.142 "enable_placement_id": 0, 00:29:51.142 "enable_zerocopy_send_server": true, 00:29:51.142 "enable_zerocopy_send_client": false, 00:29:51.142 "zerocopy_threshold": 0, 00:29:51.142 "tls_version": 0, 00:29:51.142 "enable_ktls": false 00:29:51.142 } 00:29:51.142 } 00:29:51.142 ] 00:29:51.142 }, 00:29:51.142 { 00:29:51.142 "subsystem": "vmd", 00:29:51.142 "config": [] 00:29:51.142 }, 00:29:51.142 { 00:29:51.142 "subsystem": "accel", 00:29:51.142 "config": [ 00:29:51.142 { 00:29:51.142 "method": "accel_set_options", 00:29:51.142 "params": { 00:29:51.142 "small_cache_size": 128, 00:29:51.142 "large_cache_size": 16, 00:29:51.142 "task_count": 2048, 00:29:51.142 "sequence_count": 2048, 00:29:51.142 "buf_count": 2048 00:29:51.142 } 00:29:51.142 } 00:29:51.142 ] 00:29:51.142 }, 00:29:51.142 { 00:29:51.142 "subsystem": "bdev", 00:29:51.142 "config": [ 00:29:51.142 { 00:29:51.142 "method": "bdev_set_options", 00:29:51.142 "params": { 00:29:51.142 "bdev_io_pool_size": 65535, 00:29:51.142 "bdev_io_cache_size": 256, 00:29:51.142 "bdev_auto_examine": true, 00:29:51.142 "iobuf_small_cache_size": 128, 00:29:51.142 "iobuf_large_cache_size": 16 00:29:51.142 } 00:29:51.143 }, 00:29:51.143 { 00:29:51.143 "method": "bdev_raid_set_options", 00:29:51.143 "params": { 00:29:51.143 "process_window_size_kb": 1024 00:29:51.143 } 00:29:51.143 }, 00:29:51.143 { 00:29:51.143 "method": "bdev_iscsi_set_options", 00:29:51.143 "params": { 00:29:51.143 "timeout_sec": 30 00:29:51.143 } 00:29:51.143 }, 00:29:51.143 { 00:29:51.143 "method": "bdev_nvme_set_options", 00:29:51.143 "params": { 00:29:51.143 "action_on_timeout": "none", 00:29:51.143 "timeout_us": 0, 00:29:51.143 "timeout_admin_us": 0, 00:29:51.143 "keep_alive_timeout_ms": 10000, 00:29:51.143 "arbitration_burst": 0, 00:29:51.143 "low_priority_weight": 0, 00:29:51.143 "medium_priority_weight": 0, 00:29:51.143 "high_priority_weight": 0, 00:29:51.143 "nvme_adminq_poll_period_us": 10000, 00:29:51.143 "nvme_ioq_poll_period_us": 0, 00:29:51.143 "io_queue_requests": 512, 00:29:51.143 "delay_cmd_submit": true, 00:29:51.143 "transport_retry_count": 4, 00:29:51.143 "bdev_retry_count": 3, 00:29:51.143 "transport_ack_timeout": 0, 00:29:51.143 "ctrlr_loss_timeout_sec": 0, 00:29:51.143 "reconnect_delay_sec": 0, 00:29:51.143 "fast_io_fail_timeout_sec": 0, 00:29:51.143 "disable_auto_failback": false, 00:29:51.143 "generate_uuids": false, 00:29:51.143 "transport_tos": 0, 00:29:51.143 "nvme_error_stat": false, 00:29:51.143 "rdma_srq_size": 0, 00:29:51.143 "io_path_stat": false, 00:29:51.143 "allow_accel_sequence": false, 00:29:51.143 "rdma_max_cq_size": 0, 00:29:51.143 "rdma_cm_event_timeout_ms": 0, 00:29:51.143 "dhchap_digests": [ 00:29:51.143 "sha256", 00:29:51.143 "sha384", 00:29:51.143 "sha512" 00:29:51.143 ], 00:29:51.143 "dhchap_dhgroups": [ 00:29:51.143 "null", 00:29:51.143 "ffdhe2048", 00:29:51.143 "ffdhe3072", 00:29:51.143 "ffdhe4096", 00:29:51.143 "ffdhe6144", 00:29:51.143 "ffdhe8192" 00:29:51.143 ] 00:29:51.143 } 00:29:51.143 }, 00:29:51.143 { 00:29:51.143 "method": "bdev_nvme_attach_controller", 00:29:51.143 "params": { 00:29:51.143 "name": "nvme0", 00:29:51.143 "trtype": "TCP", 00:29:51.143 "adrfam": "IPv4", 00:29:51.143 "traddr": "127.0.0.1", 00:29:51.143 "trsvcid": "4420", 00:29:51.143 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:29:51.143 "prchk_reftag": false, 00:29:51.143 "prchk_guard": false, 00:29:51.143 "ctrlr_loss_timeout_sec": 0, 00:29:51.143 "reconnect_delay_sec": 0, 00:29:51.143 "fast_io_fail_timeout_sec": 0, 00:29:51.143 "psk": "key0", 00:29:51.143 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:29:51.143 "hdgst": false, 00:29:51.143 "ddgst": false 00:29:51.143 } 00:29:51.143 }, 00:29:51.143 { 00:29:51.143 "method": "bdev_nvme_set_hotplug", 00:29:51.143 "params": { 00:29:51.143 "period_us": 100000, 00:29:51.143 "enable": false 00:29:51.143 } 00:29:51.143 }, 00:29:51.143 { 00:29:51.143 "method": "bdev_wait_for_examine" 00:29:51.143 } 00:29:51.143 ] 00:29:51.143 }, 00:29:51.143 { 00:29:51.143 "subsystem": "nbd", 00:29:51.143 "config": [] 00:29:51.143 } 00:29:51.143 ] 00:29:51.143 }' 00:29:51.143 18:15:44 keyring_file -- keyring/file.sh@114 -- # killprocess 786467 00:29:51.143 18:15:44 keyring_file -- common/autotest_common.sh@948 -- # '[' -z 786467 ']' 00:29:51.143 18:15:44 keyring_file -- common/autotest_common.sh@952 -- # kill -0 786467 00:29:51.143 18:15:44 keyring_file -- common/autotest_common.sh@953 -- # uname 00:29:51.143 18:15:44 keyring_file -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:29:51.143 18:15:44 keyring_file -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 786467 00:29:51.143 18:15:44 keyring_file -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:29:51.143 18:15:44 keyring_file -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:29:51.143 18:15:44 keyring_file -- common/autotest_common.sh@966 -- # echo 'killing process with pid 786467' 00:29:51.143 killing process with pid 786467 00:29:51.143 18:15:44 keyring_file -- common/autotest_common.sh@967 -- # kill 786467 00:29:51.143 Received shutdown signal, test time was about 1.000000 seconds 00:29:51.143 00:29:51.143 Latency(us) 00:29:51.143 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:51.143 =================================================================================================================== 00:29:51.143 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:29:51.143 18:15:44 keyring_file -- common/autotest_common.sh@972 -- # wait 786467 00:29:51.402 18:15:44 keyring_file -- keyring/file.sh@117 -- # bperfpid=787993 00:29:51.402 18:15:44 keyring_file -- keyring/file.sh@119 -- # waitforlisten 787993 /var/tmp/bperf.sock 00:29:51.402 18:15:44 keyring_file -- common/autotest_common.sh@829 -- # '[' -z 787993 ']' 00:29:51.402 18:15:44 keyring_file -- keyring/file.sh@115 -- # echo '{ 00:29:51.402 "subsystems": [ 00:29:51.402 { 00:29:51.402 "subsystem": "keyring", 00:29:51.402 "config": [ 00:29:51.402 { 00:29:51.402 "method": "keyring_file_add_key", 00:29:51.402 "params": { 00:29:51.402 "name": "key0", 00:29:51.402 "path": "/tmp/tmp.xPkaUTaC9A" 00:29:51.402 } 00:29:51.402 }, 00:29:51.402 { 00:29:51.402 "method": "keyring_file_add_key", 00:29:51.402 "params": { 00:29:51.402 "name": "key1", 00:29:51.402 "path": "/tmp/tmp.REXsPcHwDh" 00:29:51.402 } 00:29:51.402 } 00:29:51.402 ] 00:29:51.402 }, 00:29:51.402 { 00:29:51.402 "subsystem": "iobuf", 00:29:51.402 "config": [ 00:29:51.402 { 00:29:51.402 "method": "iobuf_set_options", 00:29:51.402 "params": { 00:29:51.402 "small_pool_count": 8192, 00:29:51.402 "large_pool_count": 1024, 00:29:51.402 "small_bufsize": 8192, 00:29:51.402 "large_bufsize": 135168 00:29:51.402 } 00:29:51.402 } 00:29:51.402 ] 00:29:51.402 }, 00:29:51.402 { 00:29:51.402 "subsystem": "sock", 00:29:51.402 "config": [ 00:29:51.402 { 00:29:51.402 "method": "sock_set_default_impl", 00:29:51.402 "params": { 00:29:51.402 "impl_name": "posix" 00:29:51.402 } 00:29:51.402 }, 00:29:51.402 { 00:29:51.402 "method": "sock_impl_set_options", 00:29:51.402 "params": { 00:29:51.402 "impl_name": "ssl", 00:29:51.402 "recv_buf_size": 4096, 00:29:51.402 "send_buf_size": 4096, 00:29:51.402 "enable_recv_pipe": true, 00:29:51.402 "enable_quickack": false, 00:29:51.402 "enable_placement_id": 0, 00:29:51.402 "enable_zerocopy_send_server": true, 00:29:51.402 "enable_zerocopy_send_client": false, 00:29:51.402 "zerocopy_threshold": 0, 00:29:51.402 "tls_version": 0, 00:29:51.402 "enable_ktls": false 00:29:51.402 } 00:29:51.402 }, 00:29:51.402 { 00:29:51.402 "method": "sock_impl_set_options", 00:29:51.402 "params": { 00:29:51.402 "impl_name": "posix", 00:29:51.402 "recv_buf_size": 2097152, 00:29:51.402 "send_buf_size": 2097152, 00:29:51.402 "enable_recv_pipe": true, 00:29:51.402 "enable_quickack": false, 00:29:51.402 "enable_placement_id": 0, 00:29:51.402 "enable_zerocopy_send_server": true, 00:29:51.402 "enable_zerocopy_send_client": false, 00:29:51.402 "zerocopy_threshold": 0, 00:29:51.402 "tls_version": 0, 00:29:51.402 "enable_ktls": false 00:29:51.402 } 00:29:51.402 } 00:29:51.402 ] 00:29:51.402 }, 00:29:51.402 { 00:29:51.402 "subsystem": "vmd", 00:29:51.402 "config": [] 00:29:51.402 }, 00:29:51.402 { 00:29:51.402 "subsystem": "accel", 00:29:51.402 "config": [ 00:29:51.402 { 00:29:51.402 "method": "accel_set_options", 00:29:51.402 "params": { 00:29:51.402 "small_cache_size": 128, 00:29:51.402 "large_cache_size": 16, 00:29:51.402 "task_count": 2048, 00:29:51.402 "sequence_count": 2048, 00:29:51.402 "buf_count": 2048 00:29:51.402 } 00:29:51.402 } 00:29:51.402 ] 00:29:51.402 }, 00:29:51.402 { 00:29:51.402 "subsystem": "bdev", 00:29:51.402 "config": [ 00:29:51.402 { 00:29:51.402 "method": "bdev_set_options", 00:29:51.402 "params": { 00:29:51.402 "bdev_io_pool_size": 65535, 00:29:51.402 "bdev_io_cache_size": 256, 00:29:51.402 "bdev_auto_examine": true, 00:29:51.402 "iobuf_small_cache_size": 128, 00:29:51.402 "iobuf_large_cache_size": 16 00:29:51.402 } 00:29:51.402 }, 00:29:51.402 { 00:29:51.402 "method": "bdev_raid_set_options", 00:29:51.402 "params": { 00:29:51.402 "process_window_size_kb": 1024 00:29:51.402 } 00:29:51.402 }, 00:29:51.402 { 00:29:51.402 "method": "bdev_iscsi_set_options", 00:29:51.402 "params": { 00:29:51.402 "timeout_sec": 30 00:29:51.402 } 00:29:51.402 }, 00:29:51.402 { 00:29:51.402 "method": "bdev_nvme_set_options", 00:29:51.402 "params": { 00:29:51.402 "action_on_timeout": "none", 00:29:51.402 "timeout_us": 0, 00:29:51.402 "timeout_admin_us": 0, 00:29:51.402 "keep_alive_timeout_ms": 10000, 00:29:51.402 "arbitration_burst": 0, 00:29:51.403 "low_priority_weight": 0, 00:29:51.403 "medium_priority_weight": 0, 00:29:51.403 "high_priority_weight": 0, 00:29:51.403 "nvme_adminq_poll_period_us": 10000, 00:29:51.403 "nvme_ioq_poll_period_us": 0, 00:29:51.403 "io_queue_requests": 512, 00:29:51.403 "delay_cmd_submit": true, 00:29:51.403 "transport_retry_count": 4, 00:29:51.403 "bdev_retry_count": 3, 00:29:51.403 "transport_ack_timeout": 0, 00:29:51.403 "ctrlr_loss_timeout_sec": 0, 00:29:51.403 "reconnect_delay_sec": 0, 00:29:51.403 "fast_io_fail_timeout_sec": 0, 00:29:51.403 "disable_auto_failback": false, 00:29:51.403 "generate_uuids": false, 00:29:51.403 "transport_tos": 0, 00:29:51.403 "nvme_error_stat": false, 00:29:51.403 "rdma_srq_size": 0, 00:29:51.403 "io_path_stat": false, 00:29:51.403 "allow_accel_sequence": false, 00:29:51.403 18:15:44 keyring_file -- keyring/file.sh@115 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -q 128 -o 4k -w randrw -M 50 -t 1 -m 2 -r /var/tmp/bperf.sock -z -c /dev/fd/63 00:29:51.403 "rdma_max_cq_size": 0, 00:29:51.403 "rdma_cm_event_timeout_ms": 0, 00:29:51.403 "dhchap_digests": [ 00:29:51.403 "sha256", 00:29:51.403 "sha384", 00:29:51.403 "sha512" 00:29:51.403 ], 00:29:51.403 "dhchap_dhgroups": [ 00:29:51.403 "null", 00:29:51.403 "ffdhe2048", 00:29:51.403 "ffdhe3072", 00:29:51.403 "ffdhe4096", 00:29:51.403 "ffdhe6144", 00:29:51.403 "ffdhe8192" 00:29:51.403 ] 00:29:51.403 } 00:29:51.403 }, 00:29:51.403 { 00:29:51.403 "method": "bdev_nvme_attach_controller", 00:29:51.403 "params": { 00:29:51.403 "name": "nvme0", 00:29:51.403 "trtype": "TCP", 00:29:51.403 "adrfam": "IPv4", 00:29:51.403 "traddr": "127.0.0.1", 00:29:51.403 "trsvcid": "4420", 00:29:51.403 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:29:51.403 "prchk_reftag": false, 00:29:51.403 "prchk_guard": false, 00:29:51.403 "ctrlr_loss_timeout_sec": 0, 00:29:51.403 "reconnect_delay_sec": 0, 00:29:51.403 "fast_io_fail_timeout_sec": 0, 00:29:51.403 "psk": "key0", 00:29:51.403 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:29:51.403 "hdgst": false, 00:29:51.403 "ddgst": false 00:29:51.403 } 00:29:51.403 }, 00:29:51.403 { 00:29:51.403 "method": "bdev_nvme_set_hotplug", 00:29:51.403 "params": { 00:29:51.403 "period_us": 100000, 00:29:51.403 "enable": false 00:29:51.403 } 00:29:51.403 }, 00:29:51.403 { 00:29:51.403 "method": "bdev_wait_for_examine" 00:29:51.403 } 00:29:51.403 ] 00:29:51.403 }, 00:29:51.403 { 00:29:51.403 "subsystem": "nbd", 00:29:51.403 "config": [] 00:29:51.403 } 00:29:51.403 ] 00:29:51.403 }' 00:29:51.403 18:15:44 keyring_file -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bperf.sock 00:29:51.403 18:15:44 keyring_file -- common/autotest_common.sh@834 -- # local max_retries=100 00:29:51.403 18:15:44 keyring_file -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:29:51.403 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:29:51.403 18:15:44 keyring_file -- common/autotest_common.sh@838 -- # xtrace_disable 00:29:51.403 18:15:44 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:29:51.403 [2024-07-15 18:15:44.994473] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:29:51.403 [2024-07-15 18:15:44.994523] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid787993 ] 00:29:51.403 EAL: No free 2048 kB hugepages reported on node 1 00:29:51.403 [2024-07-15 18:15:45.049132] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:51.403 [2024-07-15 18:15:45.117149] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:29:51.661 [2024-07-15 18:15:45.276463] bdev_nvme_rpc.c: 517:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:29:52.228 18:15:45 keyring_file -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:29:52.228 18:15:45 keyring_file -- common/autotest_common.sh@862 -- # return 0 00:29:52.228 18:15:45 keyring_file -- keyring/file.sh@120 -- # bperf_cmd keyring_get_keys 00:29:52.228 18:15:45 keyring_file -- keyring/file.sh@120 -- # jq length 00:29:52.228 18:15:45 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:52.486 18:15:45 keyring_file -- keyring/file.sh@120 -- # (( 2 == 2 )) 00:29:52.486 18:15:45 keyring_file -- keyring/file.sh@121 -- # get_refcnt key0 00:29:52.486 18:15:45 keyring_file -- keyring/common.sh@12 -- # get_key key0 00:29:52.486 18:15:45 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:29:52.486 18:15:45 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:52.486 18:15:45 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:29:52.486 18:15:45 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key0")' 00:29:52.486 18:15:46 keyring_file -- keyring/file.sh@121 -- # (( 2 == 2 )) 00:29:52.486 18:15:46 keyring_file -- keyring/file.sh@122 -- # get_refcnt key1 00:29:52.486 18:15:46 keyring_file -- keyring/common.sh@12 -- # get_key key1 00:29:52.486 18:15:46 keyring_file -- keyring/common.sh@12 -- # jq -r .refcnt 00:29:52.486 18:15:46 keyring_file -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:29:52.486 18:15:46 keyring_file -- keyring/common.sh@10 -- # jq '.[] | select(.name == "key1")' 00:29:52.486 18:15:46 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:52.745 18:15:46 keyring_file -- keyring/file.sh@122 -- # (( 1 == 1 )) 00:29:52.745 18:15:46 keyring_file -- keyring/file.sh@123 -- # bperf_cmd bdev_nvme_get_controllers 00:29:52.745 18:15:46 keyring_file -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_get_controllers 00:29:52.745 18:15:46 keyring_file -- keyring/file.sh@123 -- # jq -r '.[].name' 00:29:53.003 18:15:46 keyring_file -- keyring/file.sh@123 -- # [[ nvme0 == nvme0 ]] 00:29:53.003 18:15:46 keyring_file -- keyring/file.sh@1 -- # cleanup 00:29:53.003 18:15:46 keyring_file -- keyring/file.sh@19 -- # rm -f /tmp/tmp.xPkaUTaC9A /tmp/tmp.REXsPcHwDh 00:29:53.003 18:15:46 keyring_file -- keyring/file.sh@20 -- # killprocess 787993 00:29:53.003 18:15:46 keyring_file -- common/autotest_common.sh@948 -- # '[' -z 787993 ']' 00:29:53.003 18:15:46 keyring_file -- common/autotest_common.sh@952 -- # kill -0 787993 00:29:53.003 18:15:46 keyring_file -- common/autotest_common.sh@953 -- # uname 00:29:53.003 18:15:46 keyring_file -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:29:53.003 18:15:46 keyring_file -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 787993 00:29:53.003 18:15:46 keyring_file -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:29:53.003 18:15:46 keyring_file -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:29:53.003 18:15:46 keyring_file -- common/autotest_common.sh@966 -- # echo 'killing process with pid 787993' 00:29:53.003 killing process with pid 787993 00:29:53.003 18:15:46 keyring_file -- common/autotest_common.sh@967 -- # kill 787993 00:29:53.003 Received shutdown signal, test time was about 1.000000 seconds 00:29:53.003 00:29:53.003 Latency(us) 00:29:53.003 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:53.003 =================================================================================================================== 00:29:53.003 Total : 0.00 0.00 0.00 0.00 0.00 18446744073709551616.00 0.00 00:29:53.003 18:15:46 keyring_file -- common/autotest_common.sh@972 -- # wait 787993 00:29:53.003 18:15:46 keyring_file -- keyring/file.sh@21 -- # killprocess 786331 00:29:53.003 18:15:46 keyring_file -- common/autotest_common.sh@948 -- # '[' -z 786331 ']' 00:29:53.003 18:15:46 keyring_file -- common/autotest_common.sh@952 -- # kill -0 786331 00:29:53.003 18:15:46 keyring_file -- common/autotest_common.sh@953 -- # uname 00:29:53.263 18:15:46 keyring_file -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:29:53.263 18:15:46 keyring_file -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 786331 00:29:53.263 18:15:46 keyring_file -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:29:53.263 18:15:46 keyring_file -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:29:53.263 18:15:46 keyring_file -- common/autotest_common.sh@966 -- # echo 'killing process with pid 786331' 00:29:53.263 killing process with pid 786331 00:29:53.263 18:15:46 keyring_file -- common/autotest_common.sh@967 -- # kill 786331 00:29:53.263 [2024-07-15 18:15:46.772478] app.c:1024:log_deprecation_hits: *WARNING*: nvmf_tcp_psk_path: deprecation 'PSK path' scheduled for removal in v24.09 hit 1 times 00:29:53.263 18:15:46 keyring_file -- common/autotest_common.sh@972 -- # wait 786331 00:29:53.521 00:29:53.521 real 0m11.891s 00:29:53.521 user 0m28.116s 00:29:53.521 sys 0m2.701s 00:29:53.521 18:15:47 keyring_file -- common/autotest_common.sh@1124 -- # xtrace_disable 00:29:53.521 18:15:47 keyring_file -- common/autotest_common.sh@10 -- # set +x 00:29:53.521 ************************************ 00:29:53.521 END TEST keyring_file 00:29:53.521 ************************************ 00:29:53.521 18:15:47 -- common/autotest_common.sh@1142 -- # return 0 00:29:53.521 18:15:47 -- spdk/autotest.sh@296 -- # [[ y == y ]] 00:29:53.521 18:15:47 -- spdk/autotest.sh@297 -- # run_test keyring_linux /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/linux.sh 00:29:53.521 18:15:47 -- common/autotest_common.sh@1099 -- # '[' 2 -le 1 ']' 00:29:53.521 18:15:47 -- common/autotest_common.sh@1105 -- # xtrace_disable 00:29:53.521 18:15:47 -- common/autotest_common.sh@10 -- # set +x 00:29:53.521 ************************************ 00:29:53.521 START TEST keyring_linux 00:29:53.521 ************************************ 00:29:53.521 18:15:47 keyring_linux -- common/autotest_common.sh@1123 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/linux.sh 00:29:53.521 * Looking for test storage... 00:29:53.521 * Found test storage at /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring 00:29:53.521 18:15:47 keyring_linux -- keyring/linux.sh@9 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/keyring/common.sh 00:29:53.521 18:15:47 keyring_linux -- keyring/common.sh@4 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/test/nvmf/common.sh 00:29:53.521 18:15:47 keyring_linux -- nvmf/common.sh@7 -- # uname -s 00:29:53.780 18:15:47 keyring_linux -- nvmf/common.sh@7 -- # [[ Linux == FreeBSD ]] 00:29:53.780 18:15:47 keyring_linux -- nvmf/common.sh@9 -- # NVMF_PORT=4420 00:29:53.780 18:15:47 keyring_linux -- nvmf/common.sh@10 -- # NVMF_SECOND_PORT=4421 00:29:53.780 18:15:47 keyring_linux -- nvmf/common.sh@11 -- # NVMF_THIRD_PORT=4422 00:29:53.780 18:15:47 keyring_linux -- nvmf/common.sh@12 -- # NVMF_IP_PREFIX=192.168.100 00:29:53.780 18:15:47 keyring_linux -- nvmf/common.sh@13 -- # NVMF_IP_LEAST_ADDR=8 00:29:53.780 18:15:47 keyring_linux -- nvmf/common.sh@14 -- # NVMF_TCP_IP_ADDRESS=127.0.0.1 00:29:53.780 18:15:47 keyring_linux -- nvmf/common.sh@15 -- # NVMF_TRANSPORT_OPTS= 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@16 -- # NVMF_SERIAL=SPDKISFASTANDAWESOME 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@17 -- # nvme gen-hostnqn 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@17 -- # NVME_HOSTNQN=nqn.2014-08.org.nvmexpress:uuid:80aaeb9f-0274-ea11-906e-0017a4403562 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@18 -- # NVME_HOSTID=80aaeb9f-0274-ea11-906e-0017a4403562 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@19 -- # NVME_HOST=("--hostnqn=$NVME_HOSTNQN" "--hostid=$NVME_HOSTID") 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@20 -- # NVME_CONNECT='nvme connect' 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@21 -- # NET_TYPE=phy 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@22 -- # NVME_SUBNQN=nqn.2016-06.io.spdk:testnqn 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@45 -- # source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:29:53.781 18:15:47 keyring_linux -- scripts/common.sh@508 -- # [[ -e /bin/wpdk_common.sh ]] 00:29:53.781 18:15:47 keyring_linux -- scripts/common.sh@516 -- # [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:29:53.781 18:15:47 keyring_linux -- scripts/common.sh@517 -- # source /etc/opt/spdk-pkgdep/paths/export.sh 00:29:53.781 18:15:47 keyring_linux -- paths/export.sh@2 -- # PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:53.781 18:15:47 keyring_linux -- paths/export.sh@3 -- # PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:53.781 18:15:47 keyring_linux -- paths/export.sh@4 -- # PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:53.781 18:15:47 keyring_linux -- paths/export.sh@5 -- # export PATH 00:29:53.781 18:15:47 keyring_linux -- paths/export.sh@6 -- # echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@47 -- # : 0 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@48 -- # export NVMF_APP_SHM_ID 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@49 -- # build_nvmf_app_args 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@25 -- # '[' 0 -eq 1 ']' 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@29 -- # NVMF_APP+=(-i "$NVMF_APP_SHM_ID" -e 0xFFFF) 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@31 -- # NVMF_APP+=("${NO_HUGE[@]}") 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@33 -- # '[' -n '' ']' 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@35 -- # '[' 0 -eq 1 ']' 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@51 -- # have_pci_nics=0 00:29:53.781 18:15:47 keyring_linux -- keyring/common.sh@6 -- # bperfsock=/var/tmp/bperf.sock 00:29:53.781 18:15:47 keyring_linux -- keyring/linux.sh@11 -- # subnqn=nqn.2016-06.io.spdk:cnode0 00:29:53.781 18:15:47 keyring_linux -- keyring/linux.sh@12 -- # hostnqn=nqn.2016-06.io.spdk:host0 00:29:53.781 18:15:47 keyring_linux -- keyring/linux.sh@13 -- # key0=00112233445566778899aabbccddeeff 00:29:53.781 18:15:47 keyring_linux -- keyring/linux.sh@14 -- # key1=112233445566778899aabbccddeeff00 00:29:53.781 18:15:47 keyring_linux -- keyring/linux.sh@45 -- # trap cleanup EXIT 00:29:53.781 18:15:47 keyring_linux -- keyring/linux.sh@47 -- # prep_key key0 00112233445566778899aabbccddeeff 0 /tmp/:spdk-test:key0 00:29:53.781 18:15:47 keyring_linux -- keyring/common.sh@15 -- # local name key digest path 00:29:53.781 18:15:47 keyring_linux -- keyring/common.sh@17 -- # name=key0 00:29:53.781 18:15:47 keyring_linux -- keyring/common.sh@17 -- # key=00112233445566778899aabbccddeeff 00:29:53.781 18:15:47 keyring_linux -- keyring/common.sh@17 -- # digest=0 00:29:53.781 18:15:47 keyring_linux -- keyring/common.sh@18 -- # path=/tmp/:spdk-test:key0 00:29:53.781 18:15:47 keyring_linux -- keyring/common.sh@20 -- # format_interchange_psk 00112233445566778899aabbccddeeff 0 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@715 -- # format_key NVMeTLSkey-1 00112233445566778899aabbccddeeff 0 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@702 -- # local prefix key digest 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@704 -- # prefix=NVMeTLSkey-1 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@704 -- # key=00112233445566778899aabbccddeeff 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@704 -- # digest=0 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@705 -- # python - 00:29:53.781 18:15:47 keyring_linux -- keyring/common.sh@21 -- # chmod 0600 /tmp/:spdk-test:key0 00:29:53.781 18:15:47 keyring_linux -- keyring/common.sh@23 -- # echo /tmp/:spdk-test:key0 00:29:53.781 /tmp/:spdk-test:key0 00:29:53.781 18:15:47 keyring_linux -- keyring/linux.sh@48 -- # prep_key key1 112233445566778899aabbccddeeff00 0 /tmp/:spdk-test:key1 00:29:53.781 18:15:47 keyring_linux -- keyring/common.sh@15 -- # local name key digest path 00:29:53.781 18:15:47 keyring_linux -- keyring/common.sh@17 -- # name=key1 00:29:53.781 18:15:47 keyring_linux -- keyring/common.sh@17 -- # key=112233445566778899aabbccddeeff00 00:29:53.781 18:15:47 keyring_linux -- keyring/common.sh@17 -- # digest=0 00:29:53.781 18:15:47 keyring_linux -- keyring/common.sh@18 -- # path=/tmp/:spdk-test:key1 00:29:53.781 18:15:47 keyring_linux -- keyring/common.sh@20 -- # format_interchange_psk 112233445566778899aabbccddeeff00 0 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@715 -- # format_key NVMeTLSkey-1 112233445566778899aabbccddeeff00 0 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@702 -- # local prefix key digest 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@704 -- # prefix=NVMeTLSkey-1 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@704 -- # key=112233445566778899aabbccddeeff00 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@704 -- # digest=0 00:29:53.781 18:15:47 keyring_linux -- nvmf/common.sh@705 -- # python - 00:29:53.781 18:15:47 keyring_linux -- keyring/common.sh@21 -- # chmod 0600 /tmp/:spdk-test:key1 00:29:53.781 18:15:47 keyring_linux -- keyring/common.sh@23 -- # echo /tmp/:spdk-test:key1 00:29:53.781 /tmp/:spdk-test:key1 00:29:53.781 18:15:47 keyring_linux -- keyring/linux.sh@51 -- # tgtpid=788537 00:29:53.781 18:15:47 keyring_linux -- keyring/linux.sh@53 -- # waitforlisten 788537 00:29:53.781 18:15:47 keyring_linux -- keyring/linux.sh@50 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/bin/spdk_tgt 00:29:53.781 18:15:47 keyring_linux -- common/autotest_common.sh@829 -- # '[' -z 788537 ']' 00:29:53.781 18:15:47 keyring_linux -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/spdk.sock 00:29:53.781 18:15:47 keyring_linux -- common/autotest_common.sh@834 -- # local max_retries=100 00:29:53.781 18:15:47 keyring_linux -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock...' 00:29:53.781 Waiting for process to start up and listen on UNIX domain socket /var/tmp/spdk.sock... 00:29:53.781 18:15:47 keyring_linux -- common/autotest_common.sh@838 -- # xtrace_disable 00:29:53.781 18:15:47 keyring_linux -- common/autotest_common.sh@10 -- # set +x 00:29:53.781 [2024-07-15 18:15:47.412369] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:29:53.781 [2024-07-15 18:15:47.412414] [ DPDK EAL parameters: spdk_tgt --no-shconf -c 0x1 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid788537 ] 00:29:53.781 EAL: No free 2048 kB hugepages reported on node 1 00:29:53.781 [2024-07-15 18:15:47.465856] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:54.040 [2024-07-15 18:15:47.545792] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 0 00:29:54.607 18:15:48 keyring_linux -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:29:54.607 18:15:48 keyring_linux -- common/autotest_common.sh@862 -- # return 0 00:29:54.607 18:15:48 keyring_linux -- keyring/linux.sh@54 -- # rpc_cmd 00:29:54.607 18:15:48 keyring_linux -- common/autotest_common.sh@559 -- # xtrace_disable 00:29:54.607 18:15:48 keyring_linux -- common/autotest_common.sh@10 -- # set +x 00:29:54.607 [2024-07-15 18:15:48.214448] tcp.c: 701:nvmf_tcp_create: *NOTICE*: *** TCP Transport Init *** 00:29:54.607 null0 00:29:54.607 [2024-07-15 18:15:48.246491] tcp.c: 966:nvmf_tcp_listen: *NOTICE*: TLS support is considered experimental 00:29:54.607 [2024-07-15 18:15:48.246818] tcp.c:1016:nvmf_tcp_listen: *NOTICE*: *** NVMe/TCP Target Listening on 127.0.0.1 port 4420 *** 00:29:54.607 18:15:48 keyring_linux -- common/autotest_common.sh@587 -- # [[ 0 == 0 ]] 00:29:54.607 18:15:48 keyring_linux -- keyring/linux.sh@66 -- # keyctl add user :spdk-test:key0 NVMeTLSkey-1:00:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: @s 00:29:54.607 680696328 00:29:54.607 18:15:48 keyring_linux -- keyring/linux.sh@67 -- # keyctl add user :spdk-test:key1 NVMeTLSkey-1:00:MTEyMjMzNDQ1NTY2Nzc4ODk5YWFiYmNjZGRlZWZmMDA6CPcs: @s 00:29:54.607 174782373 00:29:54.607 18:15:48 keyring_linux -- keyring/linux.sh@70 -- # bperfpid=788595 00:29:54.607 18:15:48 keyring_linux -- keyring/linux.sh@72 -- # waitforlisten 788595 /var/tmp/bperf.sock 00:29:54.607 18:15:48 keyring_linux -- keyring/linux.sh@68 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/build/examples/bdevperf -q 128 -o 4k -w randread -t 1 -m 2 -r /var/tmp/bperf.sock -z --wait-for-rpc 00:29:54.607 18:15:48 keyring_linux -- common/autotest_common.sh@829 -- # '[' -z 788595 ']' 00:29:54.607 18:15:48 keyring_linux -- common/autotest_common.sh@833 -- # local rpc_addr=/var/tmp/bperf.sock 00:29:54.607 18:15:48 keyring_linux -- common/autotest_common.sh@834 -- # local max_retries=100 00:29:54.607 18:15:48 keyring_linux -- common/autotest_common.sh@836 -- # echo 'Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock...' 00:29:54.607 Waiting for process to start up and listen on UNIX domain socket /var/tmp/bperf.sock... 00:29:54.607 18:15:48 keyring_linux -- common/autotest_common.sh@838 -- # xtrace_disable 00:29:54.607 18:15:48 keyring_linux -- common/autotest_common.sh@10 -- # set +x 00:29:54.607 [2024-07-15 18:15:48.317519] Starting SPDK v24.09-pre git sha1 2f3522da7 / DPDK 24.03.0 initialization... 00:29:54.607 [2024-07-15 18:15:48.317560] [ DPDK EAL parameters: bdevperf --no-shconf -c 2 --huge-unlink --no-telemetry --log-level=lib.eal:6 --log-level=lib.cryptodev:5 --log-level=lib.power:5 --log-level=user1:6 --base-virtaddr=0x200000000000 --match-allocations --file-prefix=spdk_pid788595 ] 00:29:54.865 EAL: No free 2048 kB hugepages reported on node 1 00:29:54.865 [2024-07-15 18:15:48.369889] app.c: 909:spdk_app_start: *NOTICE*: Total cores available: 1 00:29:54.865 [2024-07-15 18:15:48.447901] reactor.c: 941:reactor_run: *NOTICE*: Reactor started on core 1 00:29:55.433 18:15:49 keyring_linux -- common/autotest_common.sh@858 -- # (( i == 0 )) 00:29:55.433 18:15:49 keyring_linux -- common/autotest_common.sh@862 -- # return 0 00:29:55.433 18:15:49 keyring_linux -- keyring/linux.sh@73 -- # bperf_cmd keyring_linux_set_options --enable 00:29:55.433 18:15:49 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_linux_set_options --enable 00:29:55.692 18:15:49 keyring_linux -- keyring/linux.sh@74 -- # bperf_cmd framework_start_init 00:29:55.692 18:15:49 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock framework_start_init 00:29:55.950 18:15:49 keyring_linux -- keyring/linux.sh@75 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key0 00:29:55.950 18:15:49 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key0 00:29:56.221 [2024-07-15 18:15:49.712213] bdev_nvme_rpc.c: 517:rpc_bdev_nvme_attach_controller: *NOTICE*: TLS support is considered experimental 00:29:56.221 nvme0n1 00:29:56.221 18:15:49 keyring_linux -- keyring/linux.sh@77 -- # check_keys 1 :spdk-test:key0 00:29:56.221 18:15:49 keyring_linux -- keyring/linux.sh@19 -- # local count=1 name=:spdk-test:key0 00:29:56.221 18:15:49 keyring_linux -- keyring/linux.sh@20 -- # local sn 00:29:56.221 18:15:49 keyring_linux -- keyring/linux.sh@22 -- # bperf_cmd keyring_get_keys 00:29:56.221 18:15:49 keyring_linux -- keyring/linux.sh@22 -- # jq length 00:29:56.221 18:15:49 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:56.532 18:15:49 keyring_linux -- keyring/linux.sh@22 -- # (( 1 == count )) 00:29:56.532 18:15:49 keyring_linux -- keyring/linux.sh@23 -- # (( count == 0 )) 00:29:56.532 18:15:49 keyring_linux -- keyring/linux.sh@25 -- # get_key :spdk-test:key0 00:29:56.532 18:15:49 keyring_linux -- keyring/linux.sh@25 -- # jq -r .sn 00:29:56.532 18:15:49 keyring_linux -- keyring/common.sh@10 -- # jq '.[] | select(.name == ":spdk-test:key0")' 00:29:56.532 18:15:49 keyring_linux -- keyring/common.sh@10 -- # bperf_cmd keyring_get_keys 00:29:56.532 18:15:49 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:56.532 18:15:50 keyring_linux -- keyring/linux.sh@25 -- # sn=680696328 00:29:56.532 18:15:50 keyring_linux -- keyring/linux.sh@26 -- # get_keysn :spdk-test:key0 00:29:56.532 18:15:50 keyring_linux -- keyring/linux.sh@16 -- # keyctl search @s user :spdk-test:key0 00:29:56.532 18:15:50 keyring_linux -- keyring/linux.sh@26 -- # [[ 680696328 == \6\8\0\6\9\6\3\2\8 ]] 00:29:56.532 18:15:50 keyring_linux -- keyring/linux.sh@27 -- # keyctl print 680696328 00:29:56.532 18:15:50 keyring_linux -- keyring/linux.sh@27 -- # [[ NVMeTLSkey-1:00:MDAxMTIyMzM0NDU1NjY3Nzg4OTlhYWJiY2NkZGVlZmZwJEiQ: == \N\V\M\e\T\L\S\k\e\y\-\1\:\0\0\:\M\D\A\x\M\T\I\y\M\z\M\0\N\D\U\1\N\j\Y\3\N\z\g\4\O\T\l\h\Y\W\J\i\Y\2\N\k\Z\G\V\l\Z\m\Z\w\J\E\i\Q\: ]] 00:29:56.532 18:15:50 keyring_linux -- keyring/linux.sh@79 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/examples/bdev/bdevperf/bdevperf.py -s /var/tmp/bperf.sock perform_tests 00:29:56.790 Running I/O for 1 seconds... 00:29:57.727 00:29:57.727 Latency(us) 00:29:57.727 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:57.727 Job: nvme0n1 (Core Mask 0x2, workload: randread, depth: 128, IO size: 4096) 00:29:57.727 nvme0n1 : 1.01 14748.40 57.61 0.00 0.00 8640.60 7180.47 19717.79 00:29:57.727 =================================================================================================================== 00:29:57.727 Total : 14748.40 57.61 0.00 0.00 8640.60 7180.47 19717.79 00:29:57.727 0 00:29:57.727 18:15:51 keyring_linux -- keyring/linux.sh@80 -- # bperf_cmd bdev_nvme_detach_controller nvme0 00:29:57.727 18:15:51 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_detach_controller nvme0 00:29:57.986 18:15:51 keyring_linux -- keyring/linux.sh@81 -- # check_keys 0 00:29:57.986 18:15:51 keyring_linux -- keyring/linux.sh@19 -- # local count=0 name= 00:29:57.986 18:15:51 keyring_linux -- keyring/linux.sh@20 -- # local sn 00:29:57.986 18:15:51 keyring_linux -- keyring/linux.sh@22 -- # bperf_cmd keyring_get_keys 00:29:57.986 18:15:51 keyring_linux -- keyring/linux.sh@22 -- # jq length 00:29:57.986 18:15:51 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock keyring_get_keys 00:29:57.986 18:15:51 keyring_linux -- keyring/linux.sh@22 -- # (( 0 == count )) 00:29:57.986 18:15:51 keyring_linux -- keyring/linux.sh@23 -- # (( count == 0 )) 00:29:57.986 18:15:51 keyring_linux -- keyring/linux.sh@23 -- # return 00:29:57.986 18:15:51 keyring_linux -- keyring/linux.sh@84 -- # NOT bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key1 00:29:57.986 18:15:51 keyring_linux -- common/autotest_common.sh@648 -- # local es=0 00:29:57.986 18:15:51 keyring_linux -- common/autotest_common.sh@650 -- # valid_exec_arg bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key1 00:29:57.986 18:15:51 keyring_linux -- common/autotest_common.sh@636 -- # local arg=bperf_cmd 00:29:57.986 18:15:51 keyring_linux -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:29:57.986 18:15:51 keyring_linux -- common/autotest_common.sh@640 -- # type -t bperf_cmd 00:29:57.986 18:15:51 keyring_linux -- common/autotest_common.sh@640 -- # case "$(type -t "$arg")" in 00:29:57.986 18:15:51 keyring_linux -- common/autotest_common.sh@651 -- # bperf_cmd bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key1 00:29:57.986 18:15:51 keyring_linux -- keyring/common.sh@8 -- # /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/rpc.py -s /var/tmp/bperf.sock bdev_nvme_attach_controller -b nvme0 -t tcp -a 127.0.0.1 -s 4420 -f ipv4 -n nqn.2016-06.io.spdk:cnode0 -q nqn.2016-06.io.spdk:host0 --psk :spdk-test:key1 00:29:58.246 [2024-07-15 18:15:51.828063] /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/include/spdk_internal/nvme_tcp.h: 428:nvme_tcp_read_data: *ERROR*: spdk_sock_recv() failed, errno 107: Transport endpoint is not connected 00:29:58.246 [2024-07-15 18:15:51.828947] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1fa7fd0 (107): Transport endpoint is not connected 00:29:58.246 [2024-07-15 18:15:51.829943] nvme_tcp.c:2185:nvme_tcp_qpair_process_completions: *ERROR*: Failed to flush tqpair=0x1fa7fd0 (9): Bad file descriptor 00:29:58.246 [2024-07-15 18:15:51.830944] nvme_ctrlr.c:4164:nvme_ctrlr_process_init: *ERROR*: [nqn.2016-06.io.spdk:cnode0] Ctrlr is in error state 00:29:58.246 [2024-07-15 18:15:51.830957] nvme.c: 708:nvme_ctrlr_poll_internal: *ERROR*: Failed to initialize SSD: 127.0.0.1 00:29:58.246 [2024-07-15 18:15:51.830964] nvme_ctrlr.c:1106:nvme_ctrlr_fail: *ERROR*: [nqn.2016-06.io.spdk:cnode0] in failed state. 00:29:58.246 request: 00:29:58.246 { 00:29:58.246 "name": "nvme0", 00:29:58.246 "trtype": "tcp", 00:29:58.246 "traddr": "127.0.0.1", 00:29:58.246 "adrfam": "ipv4", 00:29:58.246 "trsvcid": "4420", 00:29:58.246 "subnqn": "nqn.2016-06.io.spdk:cnode0", 00:29:58.246 "hostnqn": "nqn.2016-06.io.spdk:host0", 00:29:58.246 "prchk_reftag": false, 00:29:58.246 "prchk_guard": false, 00:29:58.246 "hdgst": false, 00:29:58.246 "ddgst": false, 00:29:58.246 "psk": ":spdk-test:key1", 00:29:58.246 "method": "bdev_nvme_attach_controller", 00:29:58.246 "req_id": 1 00:29:58.246 } 00:29:58.246 Got JSON-RPC error response 00:29:58.246 response: 00:29:58.246 { 00:29:58.246 "code": -5, 00:29:58.246 "message": "Input/output error" 00:29:58.246 } 00:29:58.246 18:15:51 keyring_linux -- common/autotest_common.sh@651 -- # es=1 00:29:58.246 18:15:51 keyring_linux -- common/autotest_common.sh@659 -- # (( es > 128 )) 00:29:58.246 18:15:51 keyring_linux -- common/autotest_common.sh@670 -- # [[ -n '' ]] 00:29:58.246 18:15:51 keyring_linux -- common/autotest_common.sh@675 -- # (( !es == 0 )) 00:29:58.246 18:15:51 keyring_linux -- keyring/linux.sh@1 -- # cleanup 00:29:58.246 18:15:51 keyring_linux -- keyring/linux.sh@38 -- # for key in key0 key1 00:29:58.246 18:15:51 keyring_linux -- keyring/linux.sh@39 -- # unlink_key key0 00:29:58.246 18:15:51 keyring_linux -- keyring/linux.sh@31 -- # local name=key0 sn 00:29:58.246 18:15:51 keyring_linux -- keyring/linux.sh@33 -- # get_keysn :spdk-test:key0 00:29:58.246 18:15:51 keyring_linux -- keyring/linux.sh@16 -- # keyctl search @s user :spdk-test:key0 00:29:58.246 18:15:51 keyring_linux -- keyring/linux.sh@33 -- # sn=680696328 00:29:58.246 18:15:51 keyring_linux -- keyring/linux.sh@34 -- # keyctl unlink 680696328 00:29:58.246 1 links removed 00:29:58.246 18:15:51 keyring_linux -- keyring/linux.sh@38 -- # for key in key0 key1 00:29:58.246 18:15:51 keyring_linux -- keyring/linux.sh@39 -- # unlink_key key1 00:29:58.246 18:15:51 keyring_linux -- keyring/linux.sh@31 -- # local name=key1 sn 00:29:58.246 18:15:51 keyring_linux -- keyring/linux.sh@33 -- # get_keysn :spdk-test:key1 00:29:58.246 18:15:51 keyring_linux -- keyring/linux.sh@16 -- # keyctl search @s user :spdk-test:key1 00:29:58.246 18:15:51 keyring_linux -- keyring/linux.sh@33 -- # sn=174782373 00:29:58.246 18:15:51 keyring_linux -- keyring/linux.sh@34 -- # keyctl unlink 174782373 00:29:58.246 1 links removed 00:29:58.246 18:15:51 keyring_linux -- keyring/linux.sh@41 -- # killprocess 788595 00:29:58.246 18:15:51 keyring_linux -- common/autotest_common.sh@948 -- # '[' -z 788595 ']' 00:29:58.246 18:15:51 keyring_linux -- common/autotest_common.sh@952 -- # kill -0 788595 00:29:58.246 18:15:51 keyring_linux -- common/autotest_common.sh@953 -- # uname 00:29:58.246 18:15:51 keyring_linux -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:29:58.246 18:15:51 keyring_linux -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 788595 00:29:58.246 18:15:51 keyring_linux -- common/autotest_common.sh@954 -- # process_name=reactor_1 00:29:58.246 18:15:51 keyring_linux -- common/autotest_common.sh@958 -- # '[' reactor_1 = sudo ']' 00:29:58.246 18:15:51 keyring_linux -- common/autotest_common.sh@966 -- # echo 'killing process with pid 788595' 00:29:58.246 killing process with pid 788595 00:29:58.246 18:15:51 keyring_linux -- common/autotest_common.sh@967 -- # kill 788595 00:29:58.246 Received shutdown signal, test time was about 1.000000 seconds 00:29:58.246 00:29:58.246 Latency(us) 00:29:58.246 Device Information : runtime(s) IOPS MiB/s Fail/s TO/s Average min max 00:29:58.246 =================================================================================================================== 00:29:58.246 Total : 0.00 0.00 0.00 0.00 0.00 0.00 0.00 00:29:58.246 18:15:51 keyring_linux -- common/autotest_common.sh@972 -- # wait 788595 00:29:58.505 18:15:52 keyring_linux -- keyring/linux.sh@42 -- # killprocess 788537 00:29:58.505 18:15:52 keyring_linux -- common/autotest_common.sh@948 -- # '[' -z 788537 ']' 00:29:58.505 18:15:52 keyring_linux -- common/autotest_common.sh@952 -- # kill -0 788537 00:29:58.505 18:15:52 keyring_linux -- common/autotest_common.sh@953 -- # uname 00:29:58.505 18:15:52 keyring_linux -- common/autotest_common.sh@953 -- # '[' Linux = Linux ']' 00:29:58.505 18:15:52 keyring_linux -- common/autotest_common.sh@954 -- # ps --no-headers -o comm= 788537 00:29:58.505 18:15:52 keyring_linux -- common/autotest_common.sh@954 -- # process_name=reactor_0 00:29:58.505 18:15:52 keyring_linux -- common/autotest_common.sh@958 -- # '[' reactor_0 = sudo ']' 00:29:58.505 18:15:52 keyring_linux -- common/autotest_common.sh@966 -- # echo 'killing process with pid 788537' 00:29:58.505 killing process with pid 788537 00:29:58.505 18:15:52 keyring_linux -- common/autotest_common.sh@967 -- # kill 788537 00:29:58.505 18:15:52 keyring_linux -- common/autotest_common.sh@972 -- # wait 788537 00:29:58.763 00:29:58.763 real 0m5.276s 00:29:58.763 user 0m9.231s 00:29:58.763 sys 0m1.564s 00:29:58.763 18:15:52 keyring_linux -- common/autotest_common.sh@1124 -- # xtrace_disable 00:29:58.763 18:15:52 keyring_linux -- common/autotest_common.sh@10 -- # set +x 00:29:58.763 ************************************ 00:29:58.763 END TEST keyring_linux 00:29:58.763 ************************************ 00:29:58.763 18:15:52 -- common/autotest_common.sh@1142 -- # return 0 00:29:58.763 18:15:52 -- spdk/autotest.sh@308 -- # '[' 0 -eq 1 ']' 00:29:58.763 18:15:52 -- spdk/autotest.sh@312 -- # '[' 0 -eq 1 ']' 00:29:58.763 18:15:52 -- spdk/autotest.sh@316 -- # '[' 0 -eq 1 ']' 00:29:58.763 18:15:52 -- spdk/autotest.sh@321 -- # '[' 0 -eq 1 ']' 00:29:58.763 18:15:52 -- spdk/autotest.sh@330 -- # '[' 0 -eq 1 ']' 00:29:58.763 18:15:52 -- spdk/autotest.sh@335 -- # '[' 0 -eq 1 ']' 00:29:58.763 18:15:52 -- spdk/autotest.sh@339 -- # '[' 0 -eq 1 ']' 00:29:58.763 18:15:52 -- spdk/autotest.sh@343 -- # '[' 0 -eq 1 ']' 00:29:58.763 18:15:52 -- spdk/autotest.sh@347 -- # '[' 0 -eq 1 ']' 00:29:58.763 18:15:52 -- spdk/autotest.sh@352 -- # '[' 0 -eq 1 ']' 00:29:58.763 18:15:52 -- spdk/autotest.sh@356 -- # '[' 0 -eq 1 ']' 00:29:58.763 18:15:52 -- spdk/autotest.sh@363 -- # [[ 0 -eq 1 ]] 00:29:58.763 18:15:52 -- spdk/autotest.sh@367 -- # [[ 0 -eq 1 ]] 00:29:58.763 18:15:52 -- spdk/autotest.sh@371 -- # [[ 0 -eq 1 ]] 00:29:58.763 18:15:52 -- spdk/autotest.sh@375 -- # [[ 0 -eq 1 ]] 00:29:58.763 18:15:52 -- spdk/autotest.sh@380 -- # trap - SIGINT SIGTERM EXIT 00:29:58.763 18:15:52 -- spdk/autotest.sh@382 -- # timing_enter post_cleanup 00:29:58.763 18:15:52 -- common/autotest_common.sh@722 -- # xtrace_disable 00:29:58.763 18:15:52 -- common/autotest_common.sh@10 -- # set +x 00:29:58.763 18:15:52 -- spdk/autotest.sh@383 -- # autotest_cleanup 00:29:58.763 18:15:52 -- common/autotest_common.sh@1392 -- # local autotest_es=0 00:29:58.763 18:15:52 -- common/autotest_common.sh@1393 -- # xtrace_disable 00:29:58.763 18:15:52 -- common/autotest_common.sh@10 -- # set +x 00:30:02.969 INFO: APP EXITING 00:30:02.969 INFO: killing all VMs 00:30:02.969 INFO: killing vhost app 00:30:02.969 INFO: EXIT DONE 00:30:05.503 0000:5e:00.0 (8086 0a54): Already using the nvme driver 00:30:05.503 0000:00:04.7 (8086 2021): Already using the ioatdma driver 00:30:05.503 0000:00:04.6 (8086 2021): Already using the ioatdma driver 00:30:05.503 0000:00:04.5 (8086 2021): Already using the ioatdma driver 00:30:05.503 0000:00:04.4 (8086 2021): Already using the ioatdma driver 00:30:05.503 0000:00:04.3 (8086 2021): Already using the ioatdma driver 00:30:05.503 0000:00:04.2 (8086 2021): Already using the ioatdma driver 00:30:05.503 0000:00:04.1 (8086 2021): Already using the ioatdma driver 00:30:05.503 0000:00:04.0 (8086 2021): Already using the ioatdma driver 00:30:05.503 0000:80:04.7 (8086 2021): Already using the ioatdma driver 00:30:05.503 0000:80:04.6 (8086 2021): Already using the ioatdma driver 00:30:05.503 0000:80:04.5 (8086 2021): Already using the ioatdma driver 00:30:05.761 0000:80:04.4 (8086 2021): Already using the ioatdma driver 00:30:05.761 0000:80:04.3 (8086 2021): Already using the ioatdma driver 00:30:05.761 0000:80:04.2 (8086 2021): Already using the ioatdma driver 00:30:05.761 0000:80:04.1 (8086 2021): Already using the ioatdma driver 00:30:05.761 0000:80:04.0 (8086 2021): Already using the ioatdma driver 00:30:08.292 Cleaning 00:30:08.292 Removing: /var/run/dpdk/spdk0/config 00:30:08.292 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-0 00:30:08.292 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-1 00:30:08.292 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-2 00:30:08.292 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-0-3 00:30:08.292 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-0 00:30:08.292 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-1 00:30:08.292 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-2 00:30:08.292 Removing: /var/run/dpdk/spdk0/fbarray_memseg-2048k-1-3 00:30:08.292 Removing: /var/run/dpdk/spdk0/fbarray_memzone 00:30:08.292 Removing: /var/run/dpdk/spdk0/hugepage_info 00:30:08.292 Removing: /var/run/dpdk/spdk1/config 00:30:08.292 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-0 00:30:08.292 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-1 00:30:08.292 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-2 00:30:08.292 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-0-3 00:30:08.292 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-0 00:30:08.292 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-1 00:30:08.292 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-2 00:30:08.292 Removing: /var/run/dpdk/spdk1/fbarray_memseg-2048k-1-3 00:30:08.292 Removing: /var/run/dpdk/spdk1/fbarray_memzone 00:30:08.292 Removing: /var/run/dpdk/spdk1/hugepage_info 00:30:08.292 Removing: /var/run/dpdk/spdk1/mp_socket 00:30:08.292 Removing: /var/run/dpdk/spdk2/config 00:30:08.292 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-0 00:30:08.292 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-1 00:30:08.292 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-2 00:30:08.292 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-0-3 00:30:08.292 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-0 00:30:08.292 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-1 00:30:08.292 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-2 00:30:08.292 Removing: /var/run/dpdk/spdk2/fbarray_memseg-2048k-1-3 00:30:08.292 Removing: /var/run/dpdk/spdk2/fbarray_memzone 00:30:08.292 Removing: /var/run/dpdk/spdk2/hugepage_info 00:30:08.292 Removing: /var/run/dpdk/spdk3/config 00:30:08.292 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-0 00:30:08.292 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-1 00:30:08.292 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-2 00:30:08.292 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-0-3 00:30:08.292 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-0 00:30:08.292 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-1 00:30:08.292 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-2 00:30:08.292 Removing: /var/run/dpdk/spdk3/fbarray_memseg-2048k-1-3 00:30:08.292 Removing: /var/run/dpdk/spdk3/fbarray_memzone 00:30:08.292 Removing: /var/run/dpdk/spdk3/hugepage_info 00:30:08.292 Removing: /var/run/dpdk/spdk4/config 00:30:08.551 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-0 00:30:08.551 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-1 00:30:08.551 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-2 00:30:08.551 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-0-3 00:30:08.551 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-0 00:30:08.551 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-1 00:30:08.551 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-2 00:30:08.551 Removing: /var/run/dpdk/spdk4/fbarray_memseg-2048k-1-3 00:30:08.551 Removing: /var/run/dpdk/spdk4/fbarray_memzone 00:30:08.551 Removing: /var/run/dpdk/spdk4/hugepage_info 00:30:08.551 Removing: /dev/shm/bdev_svc_trace.1 00:30:08.551 Removing: /dev/shm/nvmf_trace.0 00:30:08.551 Removing: /dev/shm/spdk_tgt_trace.pid404207 00:30:08.551 Removing: /var/run/dpdk/spdk0 00:30:08.551 Removing: /var/run/dpdk/spdk1 00:30:08.551 Removing: /var/run/dpdk/spdk2 00:30:08.551 Removing: /var/run/dpdk/spdk3 00:30:08.551 Removing: /var/run/dpdk/spdk4 00:30:08.551 Removing: /var/run/dpdk/spdk_pid402072 00:30:08.551 Removing: /var/run/dpdk/spdk_pid403134 00:30:08.551 Removing: /var/run/dpdk/spdk_pid404207 00:30:08.551 Removing: /var/run/dpdk/spdk_pid404836 00:30:08.551 Removing: /var/run/dpdk/spdk_pid405783 00:30:08.551 Removing: /var/run/dpdk/spdk_pid406005 00:30:08.551 Removing: /var/run/dpdk/spdk_pid406988 00:30:08.551 Removing: /var/run/dpdk/spdk_pid407010 00:30:08.551 Removing: /var/run/dpdk/spdk_pid407345 00:30:08.551 Removing: /var/run/dpdk/spdk_pid408850 00:30:08.551 Removing: /var/run/dpdk/spdk_pid410120 00:30:08.551 Removing: /var/run/dpdk/spdk_pid410397 00:30:08.551 Removing: /var/run/dpdk/spdk_pid410687 00:30:08.551 Removing: /var/run/dpdk/spdk_pid410991 00:30:08.551 Removing: /var/run/dpdk/spdk_pid411285 00:30:08.551 Removing: /var/run/dpdk/spdk_pid411537 00:30:08.551 Removing: /var/run/dpdk/spdk_pid411784 00:30:08.551 Removing: /var/run/dpdk/spdk_pid412062 00:30:08.551 Removing: /var/run/dpdk/spdk_pid412847 00:30:08.551 Removing: /var/run/dpdk/spdk_pid416304 00:30:08.551 Removing: /var/run/dpdk/spdk_pid416567 00:30:08.551 Removing: /var/run/dpdk/spdk_pid416833 00:30:08.551 Removing: /var/run/dpdk/spdk_pid417061 00:30:08.551 Removing: /var/run/dpdk/spdk_pid417549 00:30:08.551 Removing: /var/run/dpdk/spdk_pid417589 00:30:08.551 Removing: /var/run/dpdk/spdk_pid418058 00:30:08.551 Removing: /var/run/dpdk/spdk_pid418263 00:30:08.551 Removing: /var/run/dpdk/spdk_pid418470 00:30:08.551 Removing: /var/run/dpdk/spdk_pid418562 00:30:08.551 Removing: /var/run/dpdk/spdk_pid418818 00:30:08.551 Removing: /var/run/dpdk/spdk_pid419014 00:30:08.551 Removing: /var/run/dpdk/spdk_pid419389 00:30:08.551 Removing: /var/run/dpdk/spdk_pid419641 00:30:08.551 Removing: /var/run/dpdk/spdk_pid419929 00:30:08.551 Removing: /var/run/dpdk/spdk_pid420193 00:30:08.551 Removing: /var/run/dpdk/spdk_pid420299 00:30:08.551 Removing: /var/run/dpdk/spdk_pid420499 00:30:08.551 Removing: /var/run/dpdk/spdk_pid420755 00:30:08.551 Removing: /var/run/dpdk/spdk_pid421001 00:30:08.551 Removing: /var/run/dpdk/spdk_pid421248 00:30:08.551 Removing: /var/run/dpdk/spdk_pid421499 00:30:08.551 Removing: /var/run/dpdk/spdk_pid421749 00:30:08.551 Removing: /var/run/dpdk/spdk_pid421994 00:30:08.551 Removing: /var/run/dpdk/spdk_pid422251 00:30:08.551 Removing: /var/run/dpdk/spdk_pid422499 00:30:08.551 Removing: /var/run/dpdk/spdk_pid422753 00:30:08.551 Removing: /var/run/dpdk/spdk_pid422998 00:30:08.551 Removing: /var/run/dpdk/spdk_pid423245 00:30:08.551 Removing: /var/run/dpdk/spdk_pid423500 00:30:08.551 Removing: /var/run/dpdk/spdk_pid423747 00:30:08.551 Removing: /var/run/dpdk/spdk_pid423994 00:30:08.551 Removing: /var/run/dpdk/spdk_pid424248 00:30:08.551 Removing: /var/run/dpdk/spdk_pid424499 00:30:08.551 Removing: /var/run/dpdk/spdk_pid424747 00:30:08.551 Removing: /var/run/dpdk/spdk_pid425006 00:30:08.551 Removing: /var/run/dpdk/spdk_pid425252 00:30:08.551 Removing: /var/run/dpdk/spdk_pid425506 00:30:08.810 Removing: /var/run/dpdk/spdk_pid425728 00:30:08.810 Removing: /var/run/dpdk/spdk_pid426094 00:30:08.810 Removing: /var/run/dpdk/spdk_pid429740 00:30:08.810 Removing: /var/run/dpdk/spdk_pid473605 00:30:08.810 Removing: /var/run/dpdk/spdk_pid477915 00:30:08.810 Removing: /var/run/dpdk/spdk_pid487916 00:30:08.810 Removing: /var/run/dpdk/spdk_pid493306 00:30:08.810 Removing: /var/run/dpdk/spdk_pid497308 00:30:08.810 Removing: /var/run/dpdk/spdk_pid497988 00:30:08.810 Removing: /var/run/dpdk/spdk_pid503898 00:30:08.810 Removing: /var/run/dpdk/spdk_pid509793 00:30:08.810 Removing: /var/run/dpdk/spdk_pid509888 00:30:08.810 Removing: /var/run/dpdk/spdk_pid510824 00:30:08.810 Removing: /var/run/dpdk/spdk_pid512032 00:30:08.810 Removing: /var/run/dpdk/spdk_pid513043 00:30:08.810 Removing: /var/run/dpdk/spdk_pid513524 00:30:08.810 Removing: /var/run/dpdk/spdk_pid513719 00:30:08.810 Removing: /var/run/dpdk/spdk_pid513962 00:30:08.810 Removing: /var/run/dpdk/spdk_pid513981 00:30:08.810 Removing: /var/run/dpdk/spdk_pid513983 00:30:08.810 Removing: /var/run/dpdk/spdk_pid514895 00:30:08.810 Removing: /var/run/dpdk/spdk_pid515810 00:30:08.810 Removing: /var/run/dpdk/spdk_pid516729 00:30:08.810 Removing: /var/run/dpdk/spdk_pid517196 00:30:08.810 Removing: /var/run/dpdk/spdk_pid517201 00:30:08.810 Removing: /var/run/dpdk/spdk_pid517486 00:30:08.810 Removing: /var/run/dpdk/spdk_pid518672 00:30:08.810 Removing: /var/run/dpdk/spdk_pid519866 00:30:08.810 Removing: /var/run/dpdk/spdk_pid528208 00:30:08.810 Removing: /var/run/dpdk/spdk_pid528458 00:30:08.810 Removing: /var/run/dpdk/spdk_pid532709 00:30:08.810 Removing: /var/run/dpdk/spdk_pid538481 00:30:08.810 Removing: /var/run/dpdk/spdk_pid541162 00:30:08.810 Removing: /var/run/dpdk/spdk_pid551348 00:30:08.810 Removing: /var/run/dpdk/spdk_pid560747 00:30:08.810 Removing: /var/run/dpdk/spdk_pid562569 00:30:08.810 Removing: /var/run/dpdk/spdk_pid563509 00:30:08.810 Removing: /var/run/dpdk/spdk_pid580101 00:30:08.810 Removing: /var/run/dpdk/spdk_pid583868 00:30:08.810 Removing: /var/run/dpdk/spdk_pid609074 00:30:08.810 Removing: /var/run/dpdk/spdk_pid613548 00:30:08.810 Removing: /var/run/dpdk/spdk_pid615152 00:30:08.810 Removing: /var/run/dpdk/spdk_pid617053 00:30:08.810 Removing: /var/run/dpdk/spdk_pid617226 00:30:08.810 Removing: /var/run/dpdk/spdk_pid617470 00:30:08.810 Removing: /var/run/dpdk/spdk_pid617704 00:30:08.810 Removing: /var/run/dpdk/spdk_pid618366 00:30:08.810 Removing: /var/run/dpdk/spdk_pid620049 00:30:08.810 Removing: /var/run/dpdk/spdk_pid621034 00:30:08.810 Removing: /var/run/dpdk/spdk_pid621433 00:30:08.810 Removing: /var/run/dpdk/spdk_pid623626 00:30:08.810 Removing: /var/run/dpdk/spdk_pid624344 00:30:08.810 Removing: /var/run/dpdk/spdk_pid625069 00:30:08.810 Removing: /var/run/dpdk/spdk_pid629113 00:30:08.810 Removing: /var/run/dpdk/spdk_pid639322 00:30:08.810 Removing: /var/run/dpdk/spdk_pid643623 00:30:08.810 Removing: /var/run/dpdk/spdk_pid649377 00:30:08.810 Removing: /var/run/dpdk/spdk_pid650781 00:30:08.810 Removing: /var/run/dpdk/spdk_pid652222 00:30:08.810 Removing: /var/run/dpdk/spdk_pid656657 00:30:08.810 Removing: /var/run/dpdk/spdk_pid660763 00:30:08.810 Removing: /var/run/dpdk/spdk_pid668115 00:30:08.810 Removing: /var/run/dpdk/spdk_pid668124 00:30:08.810 Removing: /var/run/dpdk/spdk_pid672613 00:30:08.810 Removing: /var/run/dpdk/spdk_pid672837 00:30:08.810 Removing: /var/run/dpdk/spdk_pid673064 00:30:08.810 Removing: /var/run/dpdk/spdk_pid673524 00:30:08.810 Removing: /var/run/dpdk/spdk_pid673529 00:30:08.810 Removing: /var/run/dpdk/spdk_pid678006 00:30:08.810 Removing: /var/run/dpdk/spdk_pid678582 00:30:08.810 Removing: /var/run/dpdk/spdk_pid682903 00:30:08.810 Removing: /var/run/dpdk/spdk_pid685664 00:30:08.810 Removing: /var/run/dpdk/spdk_pid691568 00:30:09.068 Removing: /var/run/dpdk/spdk_pid696971 00:30:09.068 Removing: /var/run/dpdk/spdk_pid705449 00:30:09.068 Removing: /var/run/dpdk/spdk_pid712426 00:30:09.068 Removing: /var/run/dpdk/spdk_pid712428 00:30:09.068 Removing: /var/run/dpdk/spdk_pid730262 00:30:09.068 Removing: /var/run/dpdk/spdk_pid730956 00:30:09.068 Removing: /var/run/dpdk/spdk_pid731602 00:30:09.068 Removing: /var/run/dpdk/spdk_pid732154 00:30:09.068 Removing: /var/run/dpdk/spdk_pid733104 00:30:09.068 Removing: /var/run/dpdk/spdk_pid733804 00:30:09.068 Removing: /var/run/dpdk/spdk_pid734510 00:30:09.068 Removing: /var/run/dpdk/spdk_pid735101 00:30:09.068 Removing: /var/run/dpdk/spdk_pid739743 00:30:09.068 Removing: /var/run/dpdk/spdk_pid739984 00:30:09.068 Removing: /var/run/dpdk/spdk_pid745867 00:30:09.068 Removing: /var/run/dpdk/spdk_pid746092 00:30:09.068 Removing: /var/run/dpdk/spdk_pid748315 00:30:09.068 Removing: /var/run/dpdk/spdk_pid756038 00:30:09.068 Removing: /var/run/dpdk/spdk_pid756046 00:30:09.068 Removing: /var/run/dpdk/spdk_pid761063 00:30:09.068 Removing: /var/run/dpdk/spdk_pid763029 00:30:09.068 Removing: /var/run/dpdk/spdk_pid764996 00:30:09.068 Removing: /var/run/dpdk/spdk_pid766117 00:30:09.068 Removing: /var/run/dpdk/spdk_pid768227 00:30:09.068 Removing: /var/run/dpdk/spdk_pid769296 00:30:09.068 Removing: /var/run/dpdk/spdk_pid777892 00:30:09.068 Removing: /var/run/dpdk/spdk_pid778440 00:30:09.068 Removing: /var/run/dpdk/spdk_pid779063 00:30:09.068 Removing: /var/run/dpdk/spdk_pid781716 00:30:09.068 Removing: /var/run/dpdk/spdk_pid782181 00:30:09.068 Removing: /var/run/dpdk/spdk_pid782649 00:30:09.068 Removing: /var/run/dpdk/spdk_pid786331 00:30:09.068 Removing: /var/run/dpdk/spdk_pid786467 00:30:09.068 Removing: /var/run/dpdk/spdk_pid787993 00:30:09.068 Removing: /var/run/dpdk/spdk_pid788537 00:30:09.068 Removing: /var/run/dpdk/spdk_pid788595 00:30:09.068 Clean 00:30:09.068 18:16:02 -- common/autotest_common.sh@1451 -- # return 0 00:30:09.068 18:16:02 -- spdk/autotest.sh@384 -- # timing_exit post_cleanup 00:30:09.068 18:16:02 -- common/autotest_common.sh@728 -- # xtrace_disable 00:30:09.068 18:16:02 -- common/autotest_common.sh@10 -- # set +x 00:30:09.068 18:16:02 -- spdk/autotest.sh@386 -- # timing_exit autotest 00:30:09.068 18:16:02 -- common/autotest_common.sh@728 -- # xtrace_disable 00:30:09.068 18:16:02 -- common/autotest_common.sh@10 -- # set +x 00:30:09.326 18:16:02 -- spdk/autotest.sh@387 -- # chmod a+r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/timing.txt 00:30:09.326 18:16:02 -- spdk/autotest.sh@389 -- # [[ -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/udev.log ]] 00:30:09.326 18:16:02 -- spdk/autotest.sh@389 -- # rm -f /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/udev.log 00:30:09.326 18:16:02 -- spdk/autotest.sh@391 -- # hash lcov 00:30:09.326 18:16:02 -- spdk/autotest.sh@391 -- # [[ CC_TYPE=gcc == *\c\l\a\n\g* ]] 00:30:09.326 18:16:02 -- spdk/autotest.sh@393 -- # hostname 00:30:09.326 18:16:02 -- spdk/autotest.sh@393 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -c -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk -t spdk-wfp-08 -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_test.info 00:30:09.326 geninfo: WARNING: invalid characters removed from testname! 00:30:31.268 18:16:23 -- spdk/autotest.sh@394 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -a /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_base.info -a /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_test.info -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:30:32.207 18:16:25 -- spdk/autotest.sh@395 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/dpdk/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:30:34.110 18:16:27 -- spdk/autotest.sh@396 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '/usr/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:30:36.014 18:16:29 -- spdk/autotest.sh@397 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/examples/vmd/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:30:37.975 18:16:31 -- spdk/autotest.sh@398 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/app/spdk_lspci/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:30:39.351 18:16:32 -- spdk/autotest.sh@399 -- # lcov --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 --no-external -q -r /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info '*/app/spdk_top/*' -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/cov_total.info 00:30:41.255 18:16:34 -- spdk/autotest.sh@400 -- # rm -f cov_base.info cov_test.info OLD_STDOUT OLD_STDERR 00:30:41.255 18:16:34 -- common/autobuild_common.sh@15 -- $ source /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/common.sh 00:30:41.255 18:16:34 -- scripts/common.sh@508 -- $ [[ -e /bin/wpdk_common.sh ]] 00:30:41.255 18:16:34 -- scripts/common.sh@516 -- $ [[ -e /etc/opt/spdk-pkgdep/paths/export.sh ]] 00:30:41.255 18:16:34 -- scripts/common.sh@517 -- $ source /etc/opt/spdk-pkgdep/paths/export.sh 00:30:41.255 18:16:34 -- paths/export.sh@2 -- $ PATH=/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:41.255 18:16:34 -- paths/export.sh@3 -- $ PATH=/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:41.255 18:16:34 -- paths/export.sh@4 -- $ PATH=/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:41.255 18:16:34 -- paths/export.sh@5 -- $ export PATH 00:30:41.255 18:16:34 -- paths/export.sh@6 -- $ echo /opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/opt/protoc/21.7/bin:/opt/go/1.21.1/bin:/opt/golangci/1.54.2/bin:/home/sys_sgci/.local/bin:/home/sys_sgci/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/bin:/usr/local/sbin:/var/spdk/dependencies/pip/bin:/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin:/var/lib/snapd/snap/bin 00:30:41.255 18:16:34 -- common/autobuild_common.sh@443 -- $ out=/var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output 00:30:41.255 18:16:34 -- common/autobuild_common.sh@444 -- $ date +%s 00:30:41.255 18:16:34 -- common/autobuild_common.sh@444 -- $ mktemp -dt spdk_1721060194.XXXXXX 00:30:41.255 18:16:34 -- common/autobuild_common.sh@444 -- $ SPDK_WORKSPACE=/tmp/spdk_1721060194.9BX6Kk 00:30:41.255 18:16:34 -- common/autobuild_common.sh@446 -- $ [[ -n '' ]] 00:30:41.255 18:16:34 -- common/autobuild_common.sh@450 -- $ '[' -n '' ']' 00:30:41.255 18:16:34 -- common/autobuild_common.sh@453 -- $ scanbuild_exclude='--exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/' 00:30:41.255 18:16:34 -- common/autobuild_common.sh@457 -- $ scanbuild_exclude+=' --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/xnvme --exclude /tmp' 00:30:41.255 18:16:34 -- common/autobuild_common.sh@459 -- $ scanbuild='scan-build -o /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/scan-build-tmp --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/dpdk/ --exclude /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/xnvme --exclude /tmp --status-bugs' 00:30:41.255 18:16:34 -- common/autobuild_common.sh@460 -- $ get_config_params 00:30:41.255 18:16:34 -- common/autotest_common.sh@396 -- $ xtrace_disable 00:30:41.255 18:16:34 -- common/autotest_common.sh@10 -- $ set +x 00:30:41.255 18:16:34 -- common/autobuild_common.sh@460 -- $ config_params='--enable-debug --enable-werror --with-rdma --with-idxd --with-fio=/usr/src/fio --with-iscsi-initiator --disable-unit-tests --enable-ubsan --enable-coverage --with-ublk --with-vfio-user' 00:30:41.255 18:16:34 -- common/autobuild_common.sh@462 -- $ start_monitor_resources 00:30:41.255 18:16:34 -- pm/common@17 -- $ local monitor 00:30:41.255 18:16:34 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:30:41.255 18:16:34 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:30:41.255 18:16:34 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:30:41.255 18:16:34 -- pm/common@21 -- $ date +%s 00:30:41.255 18:16:34 -- pm/common@19 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:30:41.255 18:16:34 -- pm/common@21 -- $ date +%s 00:30:41.255 18:16:34 -- pm/common@25 -- $ sleep 1 00:30:41.255 18:16:34 -- pm/common@21 -- $ date +%s 00:30:41.255 18:16:34 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-vmstat -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autopackage.sh.1721060194 00:30:41.255 18:16:34 -- pm/common@21 -- $ date +%s 00:30:41.255 18:16:34 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-load -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autopackage.sh.1721060194 00:30:41.255 18:16:34 -- pm/common@21 -- $ /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-cpu-temp -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autopackage.sh.1721060194 00:30:41.255 18:16:34 -- pm/common@21 -- $ sudo -E /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/scripts/perf/pm/collect-bmc-pm -d /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power -l -p monitor.autopackage.sh.1721060194 00:30:41.255 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autopackage.sh.1721060194_collect-vmstat.pm.log 00:30:41.255 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autopackage.sh.1721060194_collect-cpu-load.pm.log 00:30:41.256 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autopackage.sh.1721060194_collect-cpu-temp.pm.log 00:30:41.256 Redirecting to /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/monitor.autopackage.sh.1721060194_collect-bmc-pm.bmc.pm.log 00:30:42.193 18:16:35 -- common/autobuild_common.sh@463 -- $ trap stop_monitor_resources EXIT 00:30:42.193 18:16:35 -- spdk/autopackage.sh@10 -- $ MAKEFLAGS=-j96 00:30:42.193 18:16:35 -- spdk/autopackage.sh@11 -- $ cd /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:30:42.193 18:16:35 -- spdk/autopackage.sh@13 -- $ [[ 0 -eq 1 ]] 00:30:42.193 18:16:35 -- spdk/autopackage.sh@18 -- $ [[ 0 -eq 0 ]] 00:30:42.193 18:16:35 -- spdk/autopackage.sh@19 -- $ timing_finish 00:30:42.193 18:16:35 -- common/autotest_common.sh@734 -- $ flamegraph=/usr/local/FlameGraph/flamegraph.pl 00:30:42.193 18:16:35 -- common/autotest_common.sh@735 -- $ '[' -x /usr/local/FlameGraph/flamegraph.pl ']' 00:30:42.193 18:16:35 -- common/autotest_common.sh@737 -- $ /usr/local/FlameGraph/flamegraph.pl --title 'Build Timing' --nametype Step: --countname seconds /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/timing.txt 00:30:42.452 18:16:35 -- spdk/autopackage.sh@20 -- $ exit 0 00:30:42.452 18:16:35 -- spdk/autopackage.sh@1 -- $ stop_monitor_resources 00:30:42.452 18:16:35 -- pm/common@29 -- $ signal_monitor_resources TERM 00:30:42.452 18:16:35 -- pm/common@40 -- $ local monitor pid pids signal=TERM 00:30:42.452 18:16:35 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:30:42.452 18:16:35 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-cpu-load.pid ]] 00:30:42.452 18:16:35 -- pm/common@44 -- $ pid=798710 00:30:42.452 18:16:35 -- pm/common@50 -- $ kill -TERM 798710 00:30:42.452 18:16:35 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:30:42.452 18:16:35 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-vmstat.pid ]] 00:30:42.452 18:16:35 -- pm/common@44 -- $ pid=798712 00:30:42.452 18:16:35 -- pm/common@50 -- $ kill -TERM 798712 00:30:42.452 18:16:35 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:30:42.452 18:16:35 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-cpu-temp.pid ]] 00:30:42.452 18:16:35 -- pm/common@44 -- $ pid=798713 00:30:42.452 18:16:35 -- pm/common@50 -- $ kill -TERM 798713 00:30:42.452 18:16:35 -- pm/common@42 -- $ for monitor in "${MONITOR_RESOURCES[@]}" 00:30:42.452 18:16:35 -- pm/common@43 -- $ [[ -e /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/collect-bmc-pm.pid ]] 00:30:42.452 18:16:35 -- pm/common@44 -- $ pid=798738 00:30:42.452 18:16:35 -- pm/common@50 -- $ sudo -E kill -TERM 798738 00:30:42.452 + [[ -n 298560 ]] 00:30:42.452 + sudo kill 298560 00:30:42.462 [Pipeline] } 00:30:42.480 [Pipeline] // stage 00:30:42.485 [Pipeline] } 00:30:42.502 [Pipeline] // timeout 00:30:42.507 [Pipeline] } 00:30:42.524 [Pipeline] // catchError 00:30:42.529 [Pipeline] } 00:30:42.546 [Pipeline] // wrap 00:30:42.551 [Pipeline] } 00:30:42.564 [Pipeline] // catchError 00:30:42.571 [Pipeline] stage 00:30:42.574 [Pipeline] { (Epilogue) 00:30:42.587 [Pipeline] catchError 00:30:42.588 [Pipeline] { 00:30:42.602 [Pipeline] echo 00:30:42.604 Cleanup processes 00:30:42.610 [Pipeline] sh 00:30:42.894 + sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:30:42.894 798833 /usr/bin/ipmitool sdr dump /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk/../output/power/sdr.cache 00:30:42.894 799112 sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:30:42.910 [Pipeline] sh 00:30:43.194 ++ sudo pgrep -af /var/jenkins/workspace/nvmf-tcp-phy-autotest/spdk 00:30:43.194 ++ grep -v 'sudo pgrep' 00:30:43.194 ++ awk '{print $1}' 00:30:43.194 + sudo kill -9 798833 00:30:43.206 [Pipeline] sh 00:30:43.490 + jbp/jenkins/jjb-config/jobs/scripts/compress_artifacts.sh 00:30:53.483 [Pipeline] sh 00:30:53.767 + jbp/jenkins/jjb-config/jobs/scripts/check_artifacts_size.sh 00:30:53.767 Artifacts sizes are good 00:30:53.781 [Pipeline] archiveArtifacts 00:30:53.788 Archiving artifacts 00:30:53.946 [Pipeline] sh 00:30:54.228 + sudo chown -R sys_sgci /var/jenkins/workspace/nvmf-tcp-phy-autotest 00:30:54.243 [Pipeline] cleanWs 00:30:54.252 [WS-CLEANUP] Deleting project workspace... 00:30:54.252 [WS-CLEANUP] Deferred wipeout is used... 00:30:54.258 [WS-CLEANUP] done 00:30:54.260 [Pipeline] } 00:30:54.278 [Pipeline] // catchError 00:30:54.291 [Pipeline] sh 00:30:54.568 + logger -p user.info -t JENKINS-CI 00:30:54.577 [Pipeline] } 00:30:54.597 [Pipeline] // stage 00:30:54.602 [Pipeline] } 00:30:54.621 [Pipeline] // node 00:30:54.626 [Pipeline] End of Pipeline 00:30:54.769 Finished: SUCCESS